Research on Graph-Retrieval Augmented Generation Based on Historical Text Knowledge Graphs
By: Yang Fan , Zhang Qi , Xing Wenqian and more
Potential Business Impact:
Helps computers understand old writings better.
This article addresses domain knowledge gaps in general large language models for historical text analysis in the context of computational humanities and AIGC technology. We propose the Graph RAG framework, combining chain-of-thought prompting, self-instruction generation, and process supervision to create a The First Four Histories character relationship dataset with minimal manual annotation. This dataset supports automated historical knowledge extraction, reducing labor costs. In the graph-augmented generation phase, we introduce a collaborative mechanism between knowledge graphs and retrieval-augmented generation, improving the alignment of general models with historical knowledge. Experiments show that the domain-specific model Xunzi-Qwen1.5-14B, with Simplified Chinese input and chain-of-thought prompting, achieves optimal performance in relation extraction (F1 = 0.68). The DeepSeek model integrated with GraphRAG improves F1 by 11% (0.08-0.19) on the open-domain C-CLUE relation extraction dataset, surpassing the F1 value of Xunzi-Qwen1.5-14B (0.12), effectively alleviating hallucinations phenomenon, and improving interpretability. This framework offers a low-resource solution for classical text knowledge extraction, advancing historical knowledge services and humanities research.
Similar Papers
A Survey of Graph Retrieval-Augmented Generation for Customized Large Language Models
Computation and Language
Helps computers understand complex topics better.
Query-Centric Graph Retrieval Augmented Generation
Computation and Language
Helps computers answer harder questions by connecting ideas.
Graph-based Approaches and Functionalities in Retrieval-Augmented Generation: A Comprehensive Survey
Information Retrieval
Helps computers answer questions using real-world facts.