Skip to content
EN DE

Hallucination Management

A lawyer files a brief with the court. The cited precedents sound convincing — case numbers, court decisions, reasoning. Except: none of these cases exist. ChatGPT fabricated them. The lawyer is sanctioned by the court.

Another case: Air Canada’s AI chatbot invents a bereavement fare policy that never existed. A tribunal rules: the airline is liable for the hallucinated policy.

Google Bard’s launch demo in 2023 contains a factual error about the James Webb Telescope. The result: roughly $100 billion drop in Alphabet’s market cap in a single day — though other market factors also contributed.

Hallucinations are not a theoretical problem. They have real, measurable consequences — financial, legal, and for your users’ trust.

LLM hallucinations are not software bugs to be fixed. They are a structural property of the architecture. LLMs generate text by predicting the most likely next token — based on statistical patterns. This means:

  • The model produces plausible-sounding text even without relevant training data
  • Output confidence does not correlate with correctness
  • The same architecture that enables creative generation also enables convincing fabrication
TypeDescriptionExample
Factual fabricationInvented facts, citations, statisticsNon-existent court cases
Entity confusionMixing attributes of real entitiesAttributing Person A’s work to Person B
Temporal errorsPresenting outdated info as current”The current CEO is…” (long since replaced)
Logical hallucinationValid-sounding but flawed reasoning chainsSeemingly valid conclusions from false premises
Source hallucinationReal-seeming but fabricated sourcesURLs, paper titles, DOIs that don’t exist

Mitigation strategies (none solves it alone)

Section titled “Mitigation strategies (none solves it alone)”

RAG (Retrieval-Augmented Generation) anchors responses in external documents. General benchmarks show reductions of 40-71%. However, for specialized domains the picture differs — a Stanford study on Legal RAG found that hallucinations ‘remain substantial, diverse, and potentially insidious,’ even with RAG. The reduction rate depends heavily on retrieval quality and domain.

Span-level verification checks each individual claim against evidence from retrieved sources. Goes beyond document-level RAG to sentence-level grounding.

Multi-candidate evaluation generates multiple responses, scores them with a factuality metric, and selects the most faithful one — without model retraining.

Human-in-the-loop is essential for high-stakes domains (healthcare, legal, finance). Doesn’t scale well, but in some contexts it’s the only responsible option.

UX patterns for hallucination-prone outputs

Section titled “UX patterns for hallucination-prone outputs”
PatternWhen to use
Source attribution (inline citations with links)Always for factual claims
Confidence indicators (visual signals)When confidence varies
”Verify this” nudgesHigh-stakes domains
Regenerate optionWhen variability is expected
Edit-in-placeProfessional / expert users
Structured output (tables over prose)When accuracy matters more than readability

The Hallucination Risk Assessment — risk and mitigation by domain:

DomainRiskRequired mitigationNice-to-have
Healthcare / Legal / FinanceCriticalHuman-in-the-loop + RAG against verified sourcesSpan-level verification
Education / ResearchHighSource attribution + verification nudgesMulti-candidate evaluation
Internal tooling / ProductivityMediumDisclaimers + regenerate option + feedback loopConfidence indicators
Creative / MarketingLowerHuman review before publicationBrand guidelines as guardrails

Measurement: Track hallucination rates by category, not just overall. Tools: RAGAS, TruLens, DeepEval.

You’re the PM of an AI legal research assistant for a mid-sized law firm. The assistant helps lawyers find relevant cases, create summaries, and suggest lines of argument.

The facts:

  • 80 lawyers use the tool daily
  • RAG system with access to a legal database of 2 million documents
  • Internal evaluation: 94% of cited sources are correct (6% hallucination rate on citations)
  • 12% of summaries contain at least one factual inaccuracy
  • One partner wants to immediately approve the tool for client-facing briefs
  • Another partner wants to shut it down entirely because “6% is unacceptable”

With 80 lawyers averaging 5 research queries per day, that’s roughly 400 queries daily — at a 6% hallucination rate, that means about 24 queries per day with fabricated sources.

How would you decide?

The best decision: Neither immediate approval nor shutdown. Keep the tool as a research aid, but with strict human-in-the-loop: every source must be verified by the lawyer before it enters a brief. In parallel, reduce the hallucination rate through span-level verification and improved retrieval quality.

Why:

  • A 6% hallucination rate on citations is too high for unsupervised use in a legal tool — the lawyer sanctions case shows the consequences
  • But 94% correct sources PLUS human verification is significantly better than manual research alone (which also has errors)
  • The Air Canada ruling shows: your company is liable for hallucinated outputs — not the model, not the provider
  • The right framing: AI as a research accelerator with human quality control, not as an autonomous legal advisor

What many get wrong: Either fixating on the hallucination rate and killing the tool (giving up real productivity gains) — or ignoring the 6% and waiting for the first court incident.

Hallucinations are not a bug you fix — they’re a risk you manage. Your job as a PM isn’t to eliminate them (you can’t), but to build products that are reliable despite hallucinations.

  • RAG reduces hallucinations by 40-71% — meaning 29-60% remain. It’s a mitigation, not a solution
  • Disclaimers are legally useful but behaviorally ineffective — users habituate within minutes. Active UX patterns (inline citations, confidence signals) work better
  • Hallucination rate improvement is logarithmic — each marginal improvement is harder. For high-stakes domains, the current rate is unacceptable regardless of trend lines

Sources: Stanford Legal RAG Study (2025), Lakera — LLM Hallucinations Guide, MDPI Hallucination Mitigation Survey, Air Canada Chatbot Tribunal Ruling (2024), arxiv Hallucination Survey (2510.24476)

Part of AI Learning — free courses from prompt to production. Jan on LinkedIn