AutoTool: Dynamic Tool Selection and Integration for Agentic Reasoning
By: Jiaru Zou , Ling Yang , Yunzhe Qi and more
Potential Business Impact:
Lets AI learn to pick the right tools.
Agentic reinforcement learning has advanced large language models (LLMs) to reason through long chain-of-thought trajectories while interleaving external tool use. Existing approaches assume a fixed inventory of tools, limiting LLM agents' adaptability to new or evolving toolsets. We present AutoTool, a framework that equips LLM agents with dynamic tool-selection capabilities throughout their reasoning trajectories. We first construct a 200k dataset with explicit tool-selection rationales across 1,000+ tools and 100+ tasks spanning mathematics, science, code generation, and multimodal reasoning. Building on this data foundation, AutoTool employs a dual-phase optimization pipeline: (i) supervised and RL-based trajectory stabilization for coherent reasoning, and (ii) KL-regularized Plackett-Luce ranking to refine consistent multi-step tool selection. Across ten diverse benchmarks, we train two base models, Qwen3-8B and Qwen2.5-VL-7B, with AutoTool. With fewer parameters, AutoTool consistently outperforms advanced LLM agents and tool-integration methods, yielding average gains of 6.4% in math & science reasoning, 4.5% in search-based QA, 7.7% in code generation, and 6.9% in multimodal understanding. In addition, AutoTool exhibits stronger generalization by dynamically leveraging unseen tools from evolving toolsets during inference.
Similar Papers
AutoTool: Efficient Tool Selection for Large Language Model Agents
Artificial Intelligence
Makes smart computer helpers work faster and cheaper.
Agentic Reasoning and Tool Integration for LLMs via Reinforcement Learning
Artificial Intelligence
Helps computers learn to use tools for harder problems.
Tool-R1: Sample-Efficient Reinforcement Learning for Agentic Tool Use
Machine Learning (CS)
Lets computers use tools to solve hard problems.