Vol.01 · No.10 CS · AI · Infra April 7, 2026

AI Glossary

GlossaryReferenceLearn
LLM & Generative AI Deep Learning AI Safety & Ethics

hallucination

Hallucination

AI hallucination is a phenomenon where a large language model or other generative AI system produces outputs that seem coherent but are factually incorrect, fabricated, or nonsensical, often presented with unjustified confidence because the model predicts likely tokens without grounding in verified facts.

Difficulty

Plain Explanation

There was a big problem with early and modern AI chatbots: they can sound very confident even when they are wrong. This concept, called a “hallucination,” explains why that happens. Think of someone who memorized the sounds of a foreign song without knowing the language. They can sing smoothly, but if you ask them what the lyrics mean, they might guess and say something that sounds right but isn’t true.

Why this happens under the hood: large language models are trained to continue text by predicting the next most likely word based on patterns they learned. During generation, they choose the next token (a small chunk of text) from a probability distribution. Decoding settings, like sampling and temperature, control how strictly the model follows high-probability choices versus exploring other options. Because the model’s objective is fluency, not fact-checking, it can produce details that are statistically plausible yet unsupported. Without an internal retrieval or verification step, if a prompt asks for precise facts, the model may fill gaps with confident but incorrect specifics.

Example & Analogy

• Legal brief gone wrong: In a 2023 New York court case, an attorney used a chatbot to help draft a filing. The brief included a citation to a legal case that did not exist because the model had generated a convincing but fabricated reference. The judge later caught the error, highlighting how severe the impact of a hallucination can be in legal contexts.

• Academic citations that never existed: In higher education, when students ask tools like ChatGPT or Gemini to cite peer‑reviewed articles, the model may output realistic‑looking titles and authors that are entirely made up. The references look credible but cannot be found in any database, misleading both writers and reviewers.

• Historical facts stated with confidence: An AI chatbot can answer questions about history in a smooth, authoritative tone yet present dates, names, or events that are wrong. Because the reply sounds polished, readers may accept the incorrect information without double‑checking.

• Visual distortion in generated images: Image‑generation systems sometimes add extra fingers or limbs to people, or insert objects that don’t belong. These outputs appear plausible at first glance, but closer inspection reveals details that are incorrect or nonsensical.

At a Glance


Text LLM HallucinationsImage Generation Hallucinations
Typical outputFluent sentences with incorrect or fabricated factsRealistic-looking images with distorted or impossible details
How users noticeCross-checking facts, failed source lookups, inconsistencies over multiple promptsVisual anomalies (extra limbs), misrendered objects, inconsistent shadows
Why it happensNext-token prediction optimizes for plausible wording, not verified truthModel synthesizes pixels from patterns; realism can override object correctness
Example prompts that trigger itRequests for precise citations, obscure facts, or ambiguous questionsComplex human poses, crowded scenes with many hands/objects
Immediate mitigation by userClarify the prompt, verify with reliable sources, avoid asking for fabricated citationsInspect details carefully, regenerate with simpler constraints, refine the description

Why It Matters

• If you accept outputs at face value, you can repeat false information with great confidence, damaging credibility and trust with your audience. • In regulated or high‑stakes contexts (like legal writing), fabricated citations or facts can lead to serious professional or legal consequences. • Product decisions based on unverified AI summaries may drift off course, wasting time and resources. • UX teams that fail to signal uncertainty or encourage verification will see user trust erode when mistakes surface later.

Where It's Used

• ChatGPT: Has been observed to generate convincing but nonexistent citations or incorrect facts when prompted for sources or obscure details. • Gemini: Similar large language model behavior can occur, such as presenting plausible but inaccurate information when asked for references. • Image generators (general): Can produce visually impressive but distorted images, like extra limbs or misplaced objects.

Curious about more?
  • Role-Specific Insights
  • What mistakes do people make?
  • How do you talk about it?
  • What should I learn next?
  • What to Read Next

Role-Specific Insights

Junior Developer: Add a verification step. If your feature surfaces facts or citations, require cross‑checks before display. Avoid prompts that ask the model to invent references; direct users to source databases instead. PM/Planner: Define when accuracy is mandatory. For those flows, set policies: no citation generation, clear uncertainty messaging, and human review for high‑stakes outputs. Track a “hallucination rate” as a product metric. Senior Engineer: Tune decoding to reduce speculative details (e.g., lower exploration). Add guardrails that ask for clarification when prompts are ambiguous, and instrument logs to detect fabricated entities and citations. Designer/Content Strategist: Signal uncertainty in the UI and encourage verification. Provide affordances for users to compare answers with trusted sources and to report suspected hallucinations quickly.

Precautions

❌ Myth: Hallucinations are just software bugs. → ✅ Reality: They arise because models predict likely words or pixels rather than checking truth; the system can be functioning as designed and still be wrong. ❌ Myth: If the AI sounds confident, it must be correct. → ✅ Reality: These systems often present misinformation confidently, and rarely signal uncertainty. ❌ Myth: Asking an AI to add citations guarantees real sources. → ✅ Reality: Models may fabricate titles, authors, and journals that look authentic but do not exist. ❌ Myth: We can simply turn off hallucinations. → ✅ Reality: Eliminating them is an open engineering challenge with current LLMs and transformers; complete prevention may not be feasible today. ❌ Myth: Any unexpected output is a hallucination. → ✅ Reality: Novel or imaginative output is appropriate in creative tasks; hallucinations are a problem when accuracy is required but the system provides falsehoods as facts.

Communication

• "Content team: the long‑form draft from the LLM included three citations that don’t exist. We need a verification step before anything ships." • "Legal flagged our chatbot’s hallucination on case law last night. For compliance, let’s prohibit citation requests and route anything legal to humans." • "Design note: users trust confident tones. Let’s add UI prompts reminding people to double‑check facts when the model answers with low‑specificity sources." • "Data QA found 9% of answers contained fabricated specifics. We should adjust decoding to be less exploratory and make follow‑up clarification the default when the LLM is uncertain." • "For the education pilot, disable reference generation. If students ask for sources, the assistant should recommend vetted databases instead of inventing citations."

Related Terms

• Large Language Model (LLM) — Predicts the next token for fluent text; powerful but prone to confident errors when facts aren’t verified. • Transformer — Core architecture enabling parallel attention across tokens; brings fluency and scale but doesn’t guarantee truthfulness. • Generative AI — Creates new text or images; great for creativity, but when tasks require accuracy, ungrounded generation can lead to hallucinations. • Computer Vision — In generation, may yield realistic yet incorrect visuals (e.g., extra limbs), a visual form of hallucination distinct from textual fabrications. • Creativity in AI — Desired in art and storytelling; unlike hallucination, it’s acceptable when novelty is requested rather than factual accuracy.

What to Read Next

  1. Large Language Model (LLM) — Understand the basic objective: next‑token prediction that drives fluent but sometimes unverified outputs.
  2. Transformer — Learn the architecture that enables LLMs to handle long, coherent text and why that doesn’t ensure factual correctness.
  3. Generative AI — See how text and image generation work across modalities, and when creative novelty is desired versus when accuracy is required.
Helpful?