Newsfeed / Glossary / Hallucination
behavior

Hallucination

Pronunciation

/həˌluːsɪˈneɪʃən/

Also known as:AI hallucinationconfabulationfactual error

What is an AI Hallucination?

An AI hallucination occurs when a large language model produces output that looks plausible but is factually wrong or unsupported by evidence. Unlike a standard software bug where a system might crash or return an error, a hallucinating LLM provides completely fabricated answers with the same tone of authority as correct ones.

Hallucinations can be categorized as:

  • Factuality errors: The model states incorrect facts
  • Faithfulness errors: The model distorts or misrepresents the source or prompt

Why Do LLMs Hallucinate?

The causes are both technical and systemic:

Technical factors:

  • Noisy or incomplete training data
  • The probabilistic nature of next-token prediction
  • Lack of grounding in external knowledge
  • Architectural limitations in memory and reasoning

Systemic incentives: OpenAI's 2025 paper "Why Language Models Hallucinate" explains that next-token prediction and benchmarks that penalize "I don't know" responses implicitly push models to bluff rather than safely refuse. Standard training rewards confident guessing over admitting uncertainty.

Hallucination Rates in 2025

The field has made significant progress:

  • Top-tier models (Gemini-2.0-Flash, o3-mini-high): 0.7% to 0.9% hallucination rates
  • Medium tier: 2% to 5%
  • Average across all models: ~9.2% for general knowledge

Four models now achieve sub-1% rates—a significant milestone for enterprise trustworthiness.

Mitigation Strategies

Retrieval-Augmented Generation (RAG): The most effective technical solution. Instead of relying on internal memory, the system retrieves relevant documents from trusted sources, then feeds this context to the LLM.

Prompt Engineering: Being specific, providing context, instructing "think step by step," and explicitly allowing "I don't know" responses.

Calibration-aware training: New approaches reward uncertainty-friendly behavior rather than confident guessing.

Human-in-the-loop: Critical decisions require human verification.

The Naming Debate

Geoffrey Hinton argues these should be called "confabulations" rather than hallucinations. Hallucination implies pathology; confabulation is normal human behavior—we construct memories rather than retrieve files. The distinction matters: confabulation suggests this is an inherent feature of generative systems, not a bug to be eliminated.

"We don't store files and retrieve them; we construct memories when we need them." — Geoffrey Hinton

The Shift in Thinking

The field has moved from "chasing zero hallucinations" to "managing uncertainty in a measurable, predictable way." Perfect accuracy may be impossible, but acceptable enterprise-level reliability is achievable.

Mentioned In

Hallucinations should be called confabulations - we do them too.

Geoffrey Hinton at 00:20:00

"Hallucinations should be called confabulations - we do them too."

Related Terms

See Also