KERAG: Knowledge-Enhanced Retrieval-Augmented Generation for Advanced Question Answering
By: Yushi Sun , Kai Sun , Yifan Ethan Xu and more
Potential Business Impact:
Helps AI answer questions more accurately using more facts.
Retrieval-Augmented Generation (RAG) mitigates hallucination in Large Language Models (LLMs) by incorporating external data, with Knowledge Graphs (KGs) offering crucial information for question answering. Traditional Knowledge Graph Question Answering (KGQA) methods rely on semantic parsing, which typically retrieves knowledge strictly necessary for answer generation, thus often suffer from low coverage due to rigid schema requirements and semantic ambiguity. We present KERAG, a novel KG-based RAG pipeline that enhances QA coverage by retrieving a broader subgraph likely to contain relevant information. Our retrieval-filtering-summarization approach, combined with fine-tuned LLMs for Chain-of-Thought reasoning on knowledge sub-graphs, reduces noises and improves QA for both simple and complex questions. Experiments demonstrate that KERAG surpasses state-of-the-art solutions by about 7% in quality and exceeds GPT-4o (Tool) by 10-21%.
Similar Papers
Knowledge Graph-extended Retrieval Augmented Generation for Question Answering
Machine Learning (CS)
AI answers questions better by using facts.
Human Cognition Inspired RAG with Knowledge Graph for Complex Problem Solving
Machine Learning (CS)
Helps computers solve hard problems by thinking step-by-step.
Graph-Enhanced Retrieval-Augmented Question Answering for E-Commerce Customer Support
Computation and Language
Answers customer questions better using product facts.