Estimating Optimal Context Length for Hybrid Retrieval-augmented Multi-document Summarization
By: Adithya Pratapa, Teruko Mitamura
Potential Business Impact:
Helps computers summarize many documents better.
Recent advances in long-context reasoning abilities of language models led to interesting applications in large-scale multi-document summarization. However, prior work has shown that these long-context models are not effective at their claimed context windows. To this end, retrieval-augmented systems provide an efficient and effective alternative. However, their performance can be highly sensitive to the choice of retrieval context length. In this work, we present a hybrid method that combines retrieval-augmented systems with long-context windows supported by recent language models. Our method first estimates the optimal retrieval length as a function of the retriever, summarizer, and dataset. On a randomly sampled subset of the dataset, we use a panel of LLMs to generate a pool of silver references. We use these silver references to estimate the optimal context length for a given RAG system configuration. Our results on the multi-document summarization task showcase the effectiveness of our method across model classes and sizes. We compare against length estimates from strong long-context benchmarks such as RULER and HELMET. Our analysis also highlights the effectiveness of our estimation method for very long-context LMs and its generalization to new classes of LMs.
Similar Papers
Large Language Models with Temporal Reasoning for Longitudinal Clinical Summarization and Prediction
Computation and Language
Helps doctors quickly understand patient history.
Can LLMs reason over extended multilingual contexts? Towards long-context evaluation beyond retrieval and haystacks
Computation and Language
Tests if computers can understand long stories.
Considering Length Diversity in Retrieval-Augmented Summarization
Computation and Language
Makes AI summaries shorter and faster.