OptiSet: Unified Optimizing Set Selection and Ranking for Retrieval-Augmented Generation
By: Yi Jiang , Sendong Zhao , Jianbo Li and more
Potential Business Impact:
Finds better information for smarter computer answers.
Retrieval-Augmented Generation (RAG) improves generation quality by incorporating evidence retrieved from large external corpora. However, most existing methods rely on statically selecting top-k passages based on individual relevance, which fails to exploit combinatorial gains among passages and often introduces substantial redundancy. To address this limitation, we propose OptiSet, a set-centric framework that unifies set selection and set-level ranking for RAG. OptiSet adopts an "Expand-then-Refine" paradigm: it first expands a query into multiple perspectives to enable a diverse candidate pool and then refines the candidate pool via re-selection to form a compact evidence set. We then devise a self-synthesis strategy without strong LLM supervision to derive preference labels from the set conditional utility changes of the generator, thereby identifying complementary and redundant evidence. Finally, we introduce a set-list wise training strategy that jointly optimizes set selection and set-level ranking, enabling the model to favor compact, high-gain evidence sets. Extensive experiments demonstrate that OptiSet improves performance on complex combinatorial problems and makes generation more efficient. The source code is publicly available.
Similar Papers
Shifting from Ranking to Set Selection for Retrieval Augmented Generation
Computation and Language
Finds better answers by picking groups of facts.
Optimizing Retrieval for RAG via Reinforced Contrastive Learning
Computation and Language
AI learns to find better information for itself.
Re-ranking the Context for Multimodal Retrieval Augmented Generation
Machine Learning (CS)
Finds better pictures for AI to answer questions.