Principled Detection of Hallucinations in Large Language Models via Multiple Testing
By: Jiawei Li, Akshayaa Magesh, Venugopal V. Veeravalli
Potential Business Impact:
Stops AI from making up wrong answers.
While Large Language Models (LLMs) have emerged as powerful foundational models to solve a variety of tasks, they have also been shown to be prone to hallucinations, i.e., generating responses that sound confident but are actually incorrect or even nonsensical. In this work, we formulate the problem of detecting hallucinations as a hypothesis testing problem and draw parallels to the problem of out-of-distribution detection in machine learning models. We propose a multiple-testing-inspired method to solve the hallucination detection problem, and provide extensive experimental results to validate the robustness of our approach against state-of-the-art methods.
Similar Papers
Principled Detection of Hallucinations in Large Language Models via Multiple Testing
Computation and Language
Stops AI from making up wrong answers.
The Illusion of Progress: Re-evaluating Hallucination Detection in LLMs
Computation and Language
Fixes AI mistakes that humans can't see.
(Im)possibility of Automated Hallucination Detection in Large Language Models
Machine Learning (CS)
Helps computers know when AI is making things up.