T1: A Tool-Oriented Conversational Dataset for Multi-Turn Agentic Planning
By: Amartya Chakraborty , Paresh Dashore , Nadia Bathaee and more
Potential Business Impact:
Helps AI plan and use tools better.
Large Language Models (LLMs) have demonstrated impressive capabilities as intelligent agents capable of solving complex problems. However, effective planning in scenarios involving dependencies between API or tool calls-particularly in multi-turn conversations-remains a significant challenge. To address this, we introduce T1, a tool-augmented, multi-domain, multi-turn conversational dataset specifically designed to capture and manage inter-tool dependencies across diverse domains. T1 enables rigorous evaluation of agents' ability to coordinate tool use across nine distinct domains (4 single domain and 5 multi-domain) with the help of an integrated caching mechanism for both short- and long-term memory, while supporting dynamic replanning-such as deciding whether to recompute or reuse cached results. Beyond facilitating research on tool use and planning, T1 also serves as a benchmark for evaluating the performance of open-source language models. We present results powered by T1-Agent, highlighting their ability to plan and reason in complex, tool-dependent scenarios.
Similar Papers
Can a Single Model Master Both Multi-turn Conversations and Tool Use? CoALM: A Unified Conversational Agentic Language Model
Artificial Intelligence
Helps chatbots understand and use tools.
TravelBench: A Real-World Benchmark for Multi-Turn and Tool-Augmented Travel Planning
Artificial Intelligence
Helps AI plan trips by talking to you.
ToolDial: Multi-turn Dialogue Generation Method for Tool-Augmented Language Models
Computation and Language
Helps computers ask questions to get better answers.