Improving the Reliability of LLMs: Combining CoT, RAG, Self-Consistency, and Self-Verification
By: Adarsh Kumar , Hwiyoon Kim , Jawahar Sai Nathani and more
Potential Business Impact:
Makes AI tell the truth, not make things up.
Hallucination, where large language models (LLMs) generate confident but incorrect or irrelevant information, remains a key limitation in their application to complex, open-ended tasks. Chain-of-thought (CoT) prompting has emerged as a promising method for improving multistep reasoning by guiding models through intermediate steps. However, CoT alone does not fully address the hallucination problem. In this work, we investigate how combining CoT with retrieval-augmented generation (RAG), as well as applying self-consistency and self-verification strategies, can reduce hallucinations and improve factual accuracy. By incorporating external knowledge sources during reasoning and enabling models to verify or revise their own outputs, we aim to generate more accurate and coherent responses. We present a comparative evaluation of baseline LLMs against CoT, CoT+RAG, self-consistency, and self-verification techniques. Our results highlight the effectiveness of each method and identify the most robust approach for minimizing hallucinations while preserving fluency and reasoning depth.
Similar Papers
Learning to Reason for Hallucination Span Detection
Computation and Language
Teaches computers to spot fake facts in writing.
Mitigating Hallucination in Large Language Models (LLMs): An Application-Oriented Survey on RAG, Reasoning, and Agentic Systems
Computation and Language
Makes AI tell the truth, not make things up.
Enhancing Factual Accuracy and Citation Generation in LLMs via Multi-Stage Self-Verification
Computation and Language
Makes AI tell the truth and show proof.