Towards Long Context Hallucination Detection
By: Siyi Liu , Kishaloy Halder , Zheng Qi and more
Potential Business Impact:
Helps computers avoid making up fake information.
Large Language Models (LLMs) have demonstrated remarkable performance across various tasks. However, they are prone to contextual hallucination, generating information that is either unsubstantiated or contradictory to the given context. Although many studies have investigated contextual hallucinations in LLMs, addressing them in long-context inputs remains an open problem. In this work, we take an initial step toward solving this problem by constructing a dataset specifically designed for long-context hallucination detection. Furthermore, we propose a novel architecture that enables pre-trained encoder models, such as BERT, to process long contexts and effectively detect contextual hallucinations through a decomposition and aggregation mechanism. Our experimental results show that the proposed architecture significantly outperforms previous models of similar size as well as LLM-based models across various metrics, while providing substantially faster inference.
Similar Papers
Real-Time Detection of Hallucinated Entities in Long-Form Generation
Computation and Language
Stops AI from making up fake facts.
Why LVLMs Are More Prone to Hallucinations in Longer Responses: The Role of Context
CV and Pattern Recognition
Stops AI from making up fake information.
How Large Language Models are Designed to Hallucinate
Computers and Society
Makes AI tell the truth, not make things up.