Graph-R1: Towards Agentic GraphRAG Framework via End-to-end Reinforcement Learning
By: Haoran Luo , Haihong E , Guanting Chen and more
Potential Business Impact:
Makes AI answers more truthful and faster.
Retrieval-Augmented Generation (RAG) mitigates hallucination in LLMs by incorporating external knowledge, but relies on chunk-based retrieval that lacks structural semantics. GraphRAG methods improve RAG by modeling knowledge as entity-relation graphs, but still face challenges in high construction cost, fixed one-time retrieval, and reliance on long-context reasoning and prompt design. To address these challenges, we propose Graph-R1, an agentic GraphRAG framework via end-to-end reinforcement learning (RL). It introduces lightweight knowledge hypergraph construction, models retrieval as a multi-turn agent-environment interaction, and optimizes the agent process via an end-to-end reward mechanism. Experiments on standard RAG datasets show that Graph-R1 outperforms traditional GraphRAG and RL-enhanced RAG methods in reasoning accuracy, retrieval efficiency, and generation quality.
Similar Papers
GraphRAG-R1: Graph Retrieval-Augmented Generation with Process-Constrained Reinforcement Learning
Machine Learning (CS)
Helps computers solve harder problems by thinking more.
RouteRAG: Efficient Retrieval-Augmented Generation from Text and Graph via Reinforcement Learning
Computation and Language
Lets computers learn from text and links.
AutoGraph-R1: End-to-End Reinforcement Learning for Knowledge Graph Construction
Computation and Language
Teaches computers to build better knowledge for answering questions.