In October 2025, the Australian government received a $440,000 report from consulting firm Deloitte — containing non-existent academic sources, fabricated citations, and a fake quote from a federal court judgement. The company submitted a revised version and issued a partial refund. The following month, Deloitte's $1.6 million health report for the Canadian government of Newfoundland and Labrador was found to contain at least four false citations to research papers that do not exist.
This is what AI hallucination looks like in the real world. Not science fiction — consulting invoices, government reports, and legal briefs contaminated by AI-generated fiction delivered with complete confidence. It's the defining reliability problem of the current AI era, and understanding it is no longer optional for anyone building with or making decisions based on AI outputs.
"Language models hallucinate because standard training and evaluation procedures reward guessing over acknowledging uncertainty." — OpenAI Research Paper, September 2025 (Kalai, Nachum, Vempala & Zhang)
The 2025 research consensus has shifted significantly. Hallucinations are no longer seen simply as a data problem or a model size problem. They are now understood as a systemic incentive problem — baked into the way models are trained and evaluated — with multiple contributing causes that interact in complex ways. Here's the full picture.
01 What Is AI Hallucination, Exactly?
The term "hallucination" refers to outputs from AI models that are fluent, confident, and plausible-sounding — but factually incorrect, unfaithful to the source, or entirely fabricated. The word is somewhat contested: some researchers prefer "confabulation" (borrowed from neuroscience, describing the brain's tendency to fill memory gaps with invented content), while others — including Rebecca Parsons, CTO of ThoughtWorks — argue that all LLM outputs are hallucinations; it's just that some of them happen to be true.
There are two main categories that matter in practice:
Factual hallucination: The model states something about the world that isn't true. A citation to a paper that doesn't exist. A statistic that was never measured. A historical event that didn't happen. A person's birthday stated with confidence despite the model never having seen it.
Faithfulness hallucination: The model's output contradicts or goes beyond its input. You give it a document to summarise and it adds facts not in the document. You ask it to answer based only on a provided context and it draws on training data instead. This category is especially dangerous in RAG (retrieval-augmented generation) systems where you assume the model is grounded in real source material.
02 The Five Root Causes
Hallucinations don't have a single cause. They're the product of multiple compounding factors — each contributing in different proportions depending on the task, the model, and the domain. Here are the five that matter most.
The causes are cumulative — a rare topic (cause 1) + accuracy-rewarding benchmark (cause 2) + probabilistic generation (cause 3) = extremely high hallucination risk. High-stakes domains like law, medicine, and science trigger all five simultaneously.
03 How Bad Is It? Real Hallucination Rates
Hallucination rates vary enormously by model, task, and domain. The headline figures from the 2025 AI Hallucination Report give the clearest picture yet — and show both how far the field has come and how far it still has to go.
Sources: AI Hallucination Report 2025 (AllAboutAI). Gemini 2.0 Flash currently holds the lowest rate at 0.7%. The o3 and o4-mini PersonQA results (33–48%) illustrate how reasoning-optimised models can trade factual reliability for reasoning performance on certain tasks.
Perhaps the most striking finding: four models now have sub-1% hallucination rates on general factual tasks — a milestone that seemed distant just two years ago. Yet on specialised tasks and rare entities, even the best models remain unreliable. Dario Amodei suggested at a 2025 Anthropic developer event that frontier models may already hallucinate less than humans on certain factual tasks — a provocative claim, but one that illustrates how the goalposts are shifting from "eliminate hallucination" to "manage uncertainty predictably."
04 Real-World Consequences
| Incident | Year | Consequence | Severity |
|---|---|---|---|
| Air Canada chatbot — fabricated bereavement refund policy, court ruled airline liable | 2024 | Legal liability established; chatbot "separate entity" defence rejected | High |
| Deloitte Australia report — $440K government report contained fake citations and fabricated court quote | 2025 | Revised report submitted; partial refund issued | High |
| Deloitte Canada — $1.6M health report had 4+ non-existent research paper citations | 2025 | Public disclosure; regulatory scrutiny of AI use in consulting | High |
| Legal filings (multiple cases) — lawyers submitted briefs citing cases that don't exist | 2023–25 | Court sanctions; professional discipline proceedings | High |
| Medical chatbots — incorrect drug dosage information in clinical assistant tools | 2024–25 | Product recalls; mandatory human review protocols introduced | High |
| ChatGPT sycophancy (4o) — model validated users' delusional beliefs; public concern raised | 2025 | OpenAI rollback; NYT coverage; RLHF sycophancy now an active research area | Medium |
The pattern is consistent: AI outputs trusted without verification in high-stakes domains. The solution is not avoiding AI — it's building verification into every workflow.
05 Every Proven Technique to Reduce Hallucination
The 2025 research consensus is clear: you cannot eliminate hallucination, but you can reduce it dramatically through a layered approach combining architectural choices, training methods, deployment patterns, and prompt engineering.
No single technique is sufficient. Production deployments in high-stakes domains should combine at minimum: RAG (grounding) + prompt instructions (uncertainty expression) + human review (verification). For lower-stakes tasks, prompt engineering alone may be sufficient.
06 The Uncomfortable Truth: Will It Ever Be Zero?
The 2025 OpenAI research paper contains a mathematically sobering result: even with perfectly clean training data, hallucination cannot be fully eliminated. The proof is elegant and depressing. A model's text generation error rate must be at least twice its sentence classification error rate. Since some questions are inherently unanswerable from training data alone — birthdays of obscure individuals, results of unrecorded meetings, details that simply never appeared in text — there will always be a floor below which hallucination rates cannot fall.
Georgia Tech's Santosh Vempala puts it plainly: "If you go to a classroom with 50 students and you know the birthdays of 49 of them, that still gives you no help with the 50th. The reality is that we won't ever get to 100% accuracy."
This mathematical floor has a practical corollary: the rarer the entity or event, the higher the hallucination risk. A model asked about Abraham Lincoln will be far more reliable than one asked about a regional politician or a mid-tier research paper. The training corpus contains millions of reliable, cross-validated sentences about well-known figures and near-zero about obscure ones. When the model encounters a question about the latter, it is working almost entirely from generalised pattern-matching with very little actual signal — and no internal mechanism to distinguish "I know this" from "I'm pattern-matching a plausible-sounding answer."
This is why the PersonQA benchmark — which specifically tests models on questions about less-famous individuals — exposes hallucination rates 10–50× higher than general knowledge benchmarks. It strips away the comfortable familiarity of well-represented training data and forces the model into exactly the territory where hallucination is most likely. The o4-mini's 48% error rate on PersonQA isn't an anomaly; it's an honest look at what happens when a model is asked about things the internet simply hasn't written much about.
The Incentive Paradox
There's a commercially awkward dimension to the hallucination problem. AI researcher Wei Xing at the University of Sheffield observed: "Fixing hallucinations would kill the product." If a model responded "I don't know" to every genuinely uncertain question, users would find it useless and seek answers elsewhere.
OpenAI's own research acknowledges this tension. The fix — penalising confident errors more than uncertainty — requires rewriting evaluation benchmarks and accepting lower scores on conventional accuracy metrics. It's the right thing to do technically. It's a harder sell commercially when customers compare benchmark numbers.
The field is moving toward calibrated uncertainty as the new standard — models that express confidence levels, flag low-certainty answers, and proactively say "I'm not certain about this — you should verify." This is less satisfying than a confident answer, and more valuable than a confident wrong one.
07 What This Means For You Right Now
AI hallucination in 2026 is a known, measurable, partially manageable risk — not a mysterious defect or a reason to abandon AI tools entirely. The correct response is the same as the correct response to any known risk: understand it, design around it, and verify where it matters.
For developers building AI applications: RAG is not optional for factual applications — it is the baseline architecture. Prompt instructions that explicitly request uncertainty expressions cost nothing and reliably help. Observability tools that log model outputs for spot-checking are a minimum viable safety net.
For end users of AI tools: treat AI outputs in specialised domains (legal, medical, financial, academic) the way you'd treat a confident colleague who might be wrong. Verify citations independently. Cross-reference statistics. Never submit AI-generated professional content without fact-checking the specific claims that matter.
For the field overall: the 2025 shift toward calibration-aware training and uncertainty-friendly evaluation is the right direction. Four models already achieving sub-1% hallucination on general tasks demonstrates the problem is tractable. The work now is domain-specific reliability, sycophancy reduction, and building the evaluation infrastructure that incentivises honesty over confident bluffing — at every level of the training pipeline.
A Practical Checklist for Reducing Hallucination Risk
If you're building or deploying AI systems today, here's a concrete checklist. For factual applications (legal research, medical information, financial analysis): implement RAG with span-level citation so every claim can be traced to a source document. Add a verification prompt after generation — "Review your response and flag any claims you are uncertain about." Enable structured output so hallucinated figures can't hide inside prose.
For summarisation tasks: always instruct the model to stay faithful to the provided text and never add external information. Follow up with a faithfulness check: ask the model to identify any sentence in its summary that goes beyond what the source document says. Studies show this two-pass approach reduces faithfulness hallucinations by 40–60% on documents longer than 2,000 words.
For conversational AI: actively test for sycophancy before deploying. Present the model with incorrect "facts" stated confidently by a simulated user and check whether it agrees or corrects. Calibrate your system prompt to reinforce correction over agreement: "If the user states something factually incorrect, politely correct them with evidence rather than agreeing." Monitor output logs for hedging language — too little hedging on ambiguous topics is a hallucination risk signal.
The single most underused technique is simply asking the model to express uncertainty. A prompt that ends with "If you are not certain of any claim in your response, say so explicitly" reliably produces more accurate, better-hedged outputs — at zero cost. The model's uncertainty is already there; you just need to give it explicit permission to surface it.