Hallucination refers to the phenomenon in which an AI model generates information that is not based on facts as if it were correct. It stems from the mechanism by which LLMs generate "plausible" text from patterns in training data, and is considered difficult to eliminate entirely.
Just as human hallucinations involve perceiving things that do not exist, AI hallucinations also generate "facts that do not exist." However, unlike human hallucinations in a fundamental sense, LLMs have no mechanism for determining whether something is factual. They simply generate the next most probable token in a chain, and the result may happen to align with reality or may be entirely fabricated.
Hallucinations take several characteristic forms. Representative examples include citations of non-existent papers (generating fictitious author names and DOIs), false biographical details attributed to real individuals, and the fabrication of plausible-sounding numerical data.
What makes this particularly troublesome is that hallucinated output is grammatically correct and blends naturally into context. Obvious errors are easy to spot, but the pattern of being "90% correct and 10% false" is what makes detection so difficult.
The most promising mitigation approach at present is the adoption of RAG (Retrieval-Augmented Generation). Before the model generates a response, it retrieves relevant information from an external knowledge base and uses that information as the basis for its answer, thereby increasing the probability of factually grounded output.
Another direction is the incorporation of HITL (Human-in-the-Loop). By designing a workflow in which humans review AI output, the risk of hallucinations making their way into final deliverables is reduced. In fields such as medicine and law, where the cost of misinformation is high, this combination is becoming the de facto standard.


A technique that cross-references LLM outputs with external data sources and search results to generate factually grounded responses. A core method for reducing hallucinations.

HITL (Human-in-the-Loop) is an approach that incorporates into the design a process by which humans review, correct, and approve the outputs of AI systems. Rather than full automation, it establishes human intervention points based on the criticality of decisions, thereby ensuring accuracy and reliability.

LLM (Large Language Model) is a general term for neural network models pre-trained on massive amounts of text data, containing billions to trillions of parameters, capable of understanding and generating natural language with high accuracy.


What is Human-in-the-Loop (HITL)? The Basics of "Human Participation" Design for Establishing AI-Driven Business Process Automation