Fine-Tuning vs. RAG for Multi-Hop Question Answering with Novel Knowledge
By: Zhuoyi Yang , Yurun Song , Iftekhar Ahmed and more
Potential Business Impact:
Helps computers answer questions using new facts.
Multi-hop question answering is widely used to evaluate the reasoning capabilities of large language models (LLMs), as it requires integrating multiple pieces of supporting knowledge to arrive at a correct answer. While prior work has explored different mechanisms for providing knowledge to LLMs, such as finetuning and retrieval-augmented generation (RAG), their relative effectiveness for multi-hop question answering remains insufficiently understood, particularly when the required knowledge is temporally novel. In this paper, we systematically compare parametric and non-parametric knowledge injection methods for open-domain multi-hop question answering. We evaluate unsupervised fine-tuning (continual pretraining), supervised fine-tuning, and retrieval-augmented generation across three 7B-parameter open-source LLMs. Experiments are conducted on two benchmarks: QASC, a standard multi-hop science question answering dataset, and a newly constructed dataset of over 10,000 multi-hop questions derived from Wikipedia events in 2024, designed to test knowledge beyond the models' pretraining cutoff. Our results show that unsupervised fine-tuning provides only limited gains over base models, suggesting that continual pretraining alone is insufficient for improving multi-hop reasoning accuracy. In contrast, retrieval-augmented generation yields substantial and consistent improvements, particularly when answering questions that rely on temporally novel information. Supervised fine-tuning achieves the highest overall accuracy across models and datasets. These findings highlight fundamental differences in how knowledge injection mechanisms support multi-hop question answering and underscore the importance of retrieval-based methods when external or compositional knowledge is required.
Similar Papers
FrugalRAG: Learning to retrieve and reason for multi-hop QA
Computation and Language
Answers questions using fewer searches.
Knowledge Compression via Question Generation: Enhancing Multihop Document Retrieval without Fine-tuning
Information Retrieval
Helps computers find answers by asking questions.
Transforming Questions and Documents for Semantically Aligned Retrieval-Augmented Generation
Computation and Language
Answers hard questions by breaking them down.