AgentLongBench: A Controllable Long Benchmark For Long-Contexts Agents via Environment Rollouts
By: Shicheng Fang , Yuxin Wang , XiaoRan Liu and more
Potential Business Impact:
Tests AI's ability to solve tricky puzzles.
The evolution of Large Language Models (LLMs) into autonomous agents necessitates the management of extensive, dynamic contexts. Current benchmarks, however, remain largely static, relying on passive retrieval tasks that fail to simulate the complexities of agent-environment interaction, such as non-linear reasoning and iterative feedback. To address this, we introduce \textbf{AgentLongBench}, which evaluates agents through simulated environment rollouts based on Lateral Thinking Puzzles. This framework generates rigorous interaction trajectories across knowledge-intensive and knowledge-free scenarios. Experiments with state-of-the-art models and memory systems (32K to 4M tokens) expose a critical weakness: while adept at static retrieval, agents struggle with the dynamic information synthesis essential for workflows. Our analysis indicates that this degradation is driven by the minimum number of tokens required to resolve a query. This factor explains why the high information density inherent in massive tool responses poses a significantly greater challenge than the memory fragmentation typical of long-turn dialogues.
Similar Papers
AgentLongBench: A Controllable Long Benchmark For Long-Contexts Agents via Environment Rollouts
Computation and Language
Tests AI's ability to solve puzzles with changing clues.
LoCoBench-Agent: An Interactive Benchmark for LLM Agents in Long-Context Software Engineering
Software Engineering
Tests AI's ability to write complex computer code.
AgencyBench: Benchmarking the Frontiers of Autonomous Agents in 1M-Token Real-World Contexts
Artificial Intelligence
Tests AI agents on real-world tasks.