A technique that cross-references LLM outputs with external data sources and search results to generate factually grounded responses. A core method for reducing hallucinations.
## What is Grounding? Grounding is a technique that cross-references LLM outputs against external, trusted data sources to generate factually accurate responses. It is positioned as a core approach for reducing hallucinations (responses that are plausible but factually incorrect). ### RAG as an Implementation Method for Grounding RAG (Retrieval-Augmented Generation) is the most common method for achieving grounding. By retrieving relevant information from external databases or documents and passing that information to the model as context, it becomes easier for the model to recognize what it does not know. However, grounding is not complete with RAG alone. If the quality of retrieved results is poor, there is a risk of grounding the model on incorrect information. ### The Multi-Layered Structure of Grounding Effective grounding is achieved through multiple layers. 1. **Retrieval Layer**: Acquires relevant information via vector search or knowledge graphs 2. **Validation Layer**: Evaluates the freshness and reliability of the retrieved information 3. **Generation Layer**: Attaches citations to responses to make them verifiable 4. **Output Layer**: Performs a final factuality check using guardrails ### Practical Effects Simply requiring the model to "respond with citations" significantly reduces the hallucination rate. However, since LLMs can also fabricate citations, it is advisable to design the system to include a post-processing mechanism that verifies whether the cited URLs actually exist.


Hallucination refers to the phenomenon in which an AI model generates information that is not based on facts as if it were correct. It stems from the mechanism by which LLMs generate "plausible" text from patterns in training data, and is considered difficult to eliminate entirely.

RAG (Retrieval-Augmented Generation) is a technique that improves the accuracy and currency of responses by retrieving relevant information from external knowledge sources and appending the results to the input of an LLM.

Multi-step reasoning is a reasoning approach in which an LLM arrives at a final answer not through a single response generation, but by going through multiple intermediate steps, such as generating sub-questions, verifying partial answers, and retrieving additional information.
