This shows you the differences between two versions of the page.
| Next revision | Previous revision | ||
| why_is_my_agent_hallucinating [2026/03/25 15:33] – Create troubleshooting guide: agent hallucination causes, diagnostics, and fixes agent | why_is_my_agent_hallucinating [2026/03/30 22:39] (current) – Restructure: footnotes as references agent | ||
|---|---|---|---|
| Line 5: | Line 5: | ||
| ===== Understanding Agent Hallucination ===== | ===== Understanding Agent Hallucination ===== | ||
| - | Unlike simple LLM hallucination, | + | Unlike simple LLM hallucination, |
| **Key statistics: | **Key statistics: | ||
| - | * Base LLMs hallucinate at least 20% on rare facts (OpenAI, 2025) | + | * Base LLMs hallucinate at least 20% on rare facts(([[https:// |
| * Clinical QA systems showed 63% hallucination rate without grounding, dropping to 1.7% with ontology grounding (Votek, 2025) | * Clinical QA systems showed 63% hallucination rate without grounding, dropping to 1.7% with ontology grounding (Votek, 2025) | ||
| * ~50% of hallucinations recur on repeated prompts; 60% resurface within 10 retries (Trends Research, 2024) | * ~50% of hallucinations recur on repeated prompts; 60% resurface within 10 retries (Trends Research, 2024) | ||
| Line 17: | Line 17: | ||
| ==== 1. Tool Result Misinterpretation ==== | ==== 1. Tool Result Misinterpretation ==== | ||
| - | Agents parse tool outputs incorrectly, | + | Agents parse tool outputs incorrectly, |
| **Symptoms: | **Symptoms: | ||
| Line 41: | Line 41: | ||
| ==== 5. Exposure Bias (Snowball Effect) ==== | ==== 5. Exposure Bias (Snowball Effect) ==== | ||
| - | Autoregressive generation means early errors cascade — each wrong token increases the probability of subsequent wrong tokens. | + | Autoregressive generation means early errors cascade — each wrong token increases the probability of subsequent wrong tokens.(([[https:// |
| **Symptoms: | **Symptoms: | ||
| Line 110: | Line 110: | ||
| ==== Fix 2: Chain-of-Verification (CoVe) ==== | ==== Fix 2: Chain-of-Verification (CoVe) ==== | ||
| - | The model drafts a response, generates verification questions, answers them independently, | + | The model drafts a response, generates verification questions, answers them independently, |
| <code python> | <code python> | ||
| Line 270: | Line 270: | ||
| print(f" | print(f" | ||
| </ | </ | ||
| - | |||
| - | ===== References ===== | ||
| - | |||
| - | * Dhuliawala et al., " | ||
| - | * Lin et al., " | ||
| - | * OpenAI, "Why Language Models Hallucinate," | ||
| - | * Oxford University, "Major Research on Hallucinating Generative Models," | ||
| - | * Stanford Digital Humanities, "Legal RAG Hallucinations," | ||
| ===== See Also ===== | ===== See Also ===== | ||
| Line 284: | Line 276: | ||
| * [[common_agent_failure_modes|Common Agent Failure Modes]] | * [[common_agent_failure_modes|Common Agent Failure Modes]] | ||
| * [[how_to_handle_rate_limits|How to Handle Rate Limits]] | * [[how_to_handle_rate_limits|How to Handle Rate Limits]] | ||
| + | |||
| + | ===== References ===== | ||