FrugalRAG: Learning to retrieve and reason for multi-hop QA
By: Abhinav Java , Srivathsan Koundinyan , Nagarajan Natarajan and more
Potential Business Impact:
Answers questions using fewer searches.
We consider the problem of answering complex questions, given access to a large unstructured document corpus. The de facto approach to solving the problem is to leverage language models that (iteratively) retrieve and reason through the retrieved documents, until the model has sufficient information to generate an answer. Attempts at improving this approach focus on retrieval-augmented generation (RAG) metrics such as accuracy and recall and can be categorized into two types: (a) fine-tuning on large question answering (QA) datasets augmented with chain-of-thought traces, and (b) leveraging RL-based fine-tuning techniques that rely on question-document relevance signals. However, efficiency in the number of retrieval searches is an equally important metric, which has received less attention. In this work, we show that: (1) Large-scale fine-tuning is not needed to improve RAG metrics, contrary to popular claims in recent literature. Specifically, a standard ReAct pipeline with improved prompts can outperform state-of-the-art methods on benchmarks such as HotPotQA. (2) Supervised and RL-based fine-tuning can help RAG from the perspective of frugality, i.e., the latency due to number of searches at inference time. For example, we show that we can achieve competitive RAG metrics at nearly half the cost (in terms of number of searches) on popular RAG benchmarks, using the same base model, and at a small training cost (1000 examples).
Similar Papers
Optimizing Retrieval Strategies for Financial Question Answering Documents in Retrieval-Augmented Generation Systems
Information Retrieval
Helps computers answer money questions better.
Investigating the Robustness of Retrieval-Augmented Generation at the Query Level
Computation and Language
Makes AI smarter by improving how it finds answers.
Transforming Questions and Documents for Semantically Aligned Retrieval-Augmented Generation
Computation and Language
Answers hard questions by breaking them down.