ToLeaP: Rethinking Development of Tool Learning with Large Language Models
By: Haotian Chen , Zijun Song , Boye Niu and more
Potential Business Impact:
Helps computers learn to use new tools better.
Tool learning, which enables large language models (LLMs) to utilize external tools effectively, has garnered increasing attention for its potential to revolutionize productivity across industries. Despite rapid development in tool learning, key challenges and opportunities remain understudied, limiting deeper insights and future advancements. In this paper, we investigate the tool learning ability of 41 prevalent LLMs by reproducing 33 benchmarks and enabling one-click evaluation for seven of them, forming a Tool Learning Platform named ToLeaP. We also collect 21 out of 33 potential training datasets to facilitate future exploration. After analyzing over 3,000 bad cases of 41 LLMs based on ToLeaP, we identify four main critical challenges: (1) benchmark limitations induce both the neglect and lack of (2) autonomous learning, (3) generalization, and (4) long-horizon task-solving capabilities of LLMs. To aid future advancements, we take a step further toward exploring potential directions, namely (1) real-world benchmark construction, (2) compatibility-aware autonomous learning, (3) rationale learning by thinking, and (4) identifying and recalling key clues. The preliminary experiments demonstrate their effectiveness, highlighting the need for further research and exploration.
Similar Papers
Alignment for Efficient Tool Calling of Large Language Models
Computation and Language
Helps computers know when to use tools.
ToolACE-DEV: Self-Improving Tool Learning via Decomposition and EVolution
Computation and Language
Lets small AI learn to use tools better.
Re-Initialization Token Learning for Tool-Augmented Large Language Models
Computation and Language
Helps computers solve math and plan better.