AI Codex
Evaluation & SafetyHow It Works

How to reduce Claude hallucinations: a practical checklist

In brief

Hallucination — Claude confidently stating something that isn't true — is the failure mode that kills trust fastest. Here's exactly how to minimize it in practice.

7 min read·Hallucination

Contents

Sign in to save

Hallucination is when Claude states something confidently that is not true — inventing a statistic, citing a source that does not exist, describing a product feature that was never real. It is the failure mode that erodes trust fastest, because the wrong answer sounds exactly like the right one.

You cannot eliminate hallucination entirely. But you can reduce it significantly with the right habits and design choices.

Why Claude halluculates in the first place

Claude is a language model — it generates text that is likely to be true, not text that has been verified against reality. When it does not know something, it does not always say so. It produces what a plausible answer would look like, based on patterns in its training data.

Hallucination is more likely when:

  • The question involves specific facts, numbers, dates, or names
  • The topic is niche, recent, or not well-represented in training data
  • The prompt implies a specific answer exists
  • The model is asked to be comprehensive rather than honest about gaps

Understanding this makes the prevention strategies make sense.

The checklist: before you prompt

Provide the source material. The most reliable way to reduce hallucination is to give Claude the information it needs rather than asking it to retrieve or infer it. Paste in the document, the data, the contract, the research paper. When Claude is working from what you give it — not from memory — accuracy goes up significantly.

Ask for uncertainty explicitly. Tell Claude it is okay to say it does not know. "If you are not confident about a specific fact, say so explicitly rather than guessing." Claude responds well to this instruction.

Narrow the scope. Instead of "tell me about the history of [company]," try "based on the following paragraph about [company], what are the key dates mentioned?" Specific questions constrained to a specific source are far less likely to produce hallucinated content.

Avoid questions that beg a specific answer. "What did the CEO of [company] say about layoffs last month?" implies that the CEO said something. If Claude does not know, it may manufacture a quote rather than disappoint you. Reframe: "Is there anything in the provided materials about [company's] recent layoffs?"

During the conversation

Ask Claude to cite its sources. "For any specific facts, tell me where you got that information." Claude will either cite a real source or flag that it is drawing from general training data. Either way, you know what to verify.

Ask the follow-up: "How confident are you in this?" After a response with specific facts, ask Claude to rate its confidence and explain where the uncertainty lies. It is often honest when asked directly.

Spot-check the surprising facts. If Claude tells you something you did not know — a statistic, a date, a specific claim — verify it before using it. Not everything, but the things that will matter if wrong.

Watch for telltale hallucination patterns:

  • Very specific numbers (89.3%, not "roughly 90%") without a source
  • References to research papers or studies — check that they actually exist
  • Quotes attributed to specific people
  • Detailed biographical information about individuals
  • Claims about current prices, policies, or recent events

Design choices that reduce hallucination

Use RAG (Retrieval-Augmented Generation) for anything knowledge-intensive. If you are building an AI application that needs to answer questions about specific products, policies, or data, connect Claude to a real knowledge base rather than relying on its training data. When Claude retrieves information from your documents before answering, it is grounded in what you have verified.

Ground Claude in the present conversation. Include the relevant facts in your prompt. "Based on the following Q3 financial data: [data]" is more reliable than "summarize our Q3 performance" when Claude has not been given the data.

Use structured prompts for high-stakes tasks. When accuracy matters, structure the prompt to force verification. "For each claim you make, mark it as either (A) taken directly from the provided document or (B) your inference. Do not include any (B) claims I have not asked for inferences."

Temperature matters for accuracy tasks. At higher temperature settings, Claude is more creative — and more likely to produce fluent-but-wrong content. For factual tasks where accuracy matters more than creativity, lower temperature (closer to 0) produces more conservative, reliable outputs. Most consumer Claude interfaces handle this automatically; it is relevant if you are using the API.

What to do when you catch a hallucination

Tell Claude. "This fact is incorrect — the actual figure is [X]. Why did you state otherwise?" Claude will typically acknowledge the error and recalibrate. It is also useful feedback for diagnosing where your prompt left too much room for fabrication.

Adjust the prompt for next time: more grounding, more explicit permission to express uncertainty, a narrower scope.

The honest baseline

No set of techniques will reduce hallucination to zero. Claude is a language model, and language models are not fact-checking machines. The right mental model is not "how do I make Claude never get things wrong" but rather "for this specific task, what level of verification do I need to apply before I rely on the output?"

For brainstorming, drafting, and thinking through problems — the cost of occasional hallucination is low. For facts in a document you will share, a contract you will sign, or a decision you will make based on the output — verify.

The failure mode is not using Claude for things that need verification. The failure mode is not verifying things that need it.

Further reading

Weekly brief

For people actually using Claude at work.

What practitioners are building, the mistakes worth avoiding, and the workflows that actually stick. No tutorials. No hype.

No spam. Unsubscribe anytime.

What to read next

Picked for where you are now

All articles →