Conflict-Aware Soft Prompting for Retrieval-Augmented Generation
By: Eunseong Choi , June Park , Hyeri Lee and more
Potential Business Impact:
Helps AI tell true facts from fake ones.
Retrieval-augmented generation (RAG) enhances the capabilities of large language models (LLMs) by incorporating external knowledge into their input prompts. However, when the retrieved context contradicts the LLM's parametric knowledge, it often fails to resolve the conflict between incorrect external context and correct parametric knowledge, known as context-memory conflict. To tackle this problem, we introduce Conflict-Aware REtrieval-Augmented Generation (CARE), consisting of a context assessor and a base LLM. The context assessor encodes compact memory token embeddings from raw context tokens. Through grounded/adversarial soft prompting, the context assessor is trained to discern unreliable context and capture a guidance signal that directs reasoning toward the more reliable knowledge source. Extensive experiments show that CARE effectively mitigates context-memory conflicts, leading to an average performance gain of 5.0\% on QA and fact-checking benchmarks, establishing a promising direction for trustworthy and adaptive RAG systems.
Similar Papers
Probing Latent Knowledge Conflict for Faithful Retrieval-Augmented Generation
Computation and Language
Makes AI answers more truthful and less wrong.
Retrieval-Augmented Generation with Conflicting Evidence
Computation and Language
AI agents debate to find true answers.
When Retrieval Succeeds and Fails: Rethinking Retrieval-Augmented Generation for LLMs
Computation and Language
Helps smart computers learn new things faster.