One Tool Is Enough: Reinforcement Learning for Repository-Level LLM Agents
By: Zhaoxi Zhang , Yitong Duan , Yanzhi Zhang and more
Locating the files and functions requiring modification in large open-source software (OSS) repositories is challenging due to their scale and structural complexity. Existing large language model (LLM)-based methods typically treat this as a repository-level retrieval task and rely on multiple auxiliary tools, which overlook code execution logic and complicate model control. We propose RepoNavigator, an LLM agent equipped with a single execution-aware tool-jumping to the definition of an invoked symbol. This unified design reflects the actual flow of code execution while simplifying tool manipulation. RepoNavigator is trained end-to-end via Reinforcement Learning (RL) directly from a pretrained model, without any closed-source distillation. Experiments demonstrate that RL-trained RepoNavigator achieves state-of-the-art performance, with the 7B model outperforming 14B baselines, the 14B model surpassing 32B competitors, and even the 32B model exceeding closed-source models such as Claude-3.7. These results confirm that integrating a single, structurally grounded tool with RL training provides an efficient and scalable solution for repository-level issue localization.
Similar Papers
Tool-integrated Reinforcement Learning for Repo Deep Search
Software Engineering
Helps computers find and fix code errors faster.
Empowering RepoQA-Agent based on Reinforcement Learning Driven by Monte-carlo Tree Search
Software Engineering
Helps computers understand and answer questions about code.
Tool-R1: Sample-Efficient Reinforcement Learning for Agentic Tool Use
Machine Learning (CS)
Lets computers use tools to solve hard problems.