Interact-RAG: Reason and Interact with the Corpus, Beyond Black-Box Retrieval
By: Yulong Hui , Chao Chen , Zhihang Fu and more
Potential Business Impact:
Lets AI ask better questions to find answers.
Retrieval-Augmented Generation (RAG) has significantly enhanced LLMs by incorporating external information. However, prevailing agentic RAG approaches are constrained by a critical limitation: they treat the retrieval process as a black-box querying operation. This confines agents' actions to query issuing, hindering its ability to tackle complex information-seeking tasks. To address this, we introduce Interact-RAG, a new paradigm that elevates the LLM agent from a passive query issuer into an active manipulator of the retrieval process. We dismantle the black-box with a Corpus Interaction Engine, equipping the agent with a set of action primitives for fine-grained control over information retrieval. To further empower the agent on the entire RAG pipeline, we first develop a reasoning-enhanced workflow, which enables both zero-shot execution and the synthesis of interaction trajectories. We then leverage this synthetic data to train a fully autonomous end-to-end agent via Supervised Fine-Tuning (SFT), followed by refinement with Reinforcement Learning (RL). Extensive experiments across six benchmarks demonstrate that Interact-RAG significantly outperforms other advanced methods, validating the efficacy of our reasoning-interaction strategy.
Similar Papers
Optimizing Retrieval for RAG via Reinforced Contrastive Learning
Computation and Language
AI learns to find better information for itself.
MARAG-R1: Beyond Single Retriever via Reinforcement-Learned Multi-Tool Agentic Retrieval
Computation and Language
Lets computers find better answers from many sources.
Test-time Corpus Feedback: From Retrieval to RAG
Information Retrieval
Lets computers ask better questions to find answers.