Hallucination
When an AI model generates information that is factually incorrect, fabricated, or not supported by its training data or provided context.
Definition
When an AI model generates information that is factually incorrect, fabricated, or not supported by its training data or provided context.
Why It Matters for AI Governance
AI hallucinations can generate fake personal data, incorrect regulatory citations, or fabricated compliance claims. Output scanning must detect hallucinated PII and factual errors before they reach end users.
How CrewCheck Handles This
CrewCheck's LLM gateway applies hallucination-related controls at the request boundary. Every AI call passes through detection, policy evaluation, and audit logging — ensuring that hallucination is addressed consistently across all teams and providers.
The governance dashboard provides real-time visibility into hallucination events, with drill-down capabilities for compliance officers and exportable evidence for auditors.