Scaling Long-Horizon LLM Agent via Context-Folding
By: Weiwei Sun , Miao Lu , Zhan Ling and more
Potential Business Impact:
Helps AI remember more for long tasks.
Large language model (LLM) agents are fundamentally constrained by context length on long-horizon tasks. We introduce Context-Folding, a framework that empowers agents to actively manage their working context. An agent can procedurally branch into a sub-trajectory to handle a subtask and then fold it upon completion, collapsing the intermediate steps while retaining a concise summary of the outcome. To make this behavior learnable, we develop an end-to-end reinforcement learning framework FoldGRPO with specific process rewards to encourage effective task decomposition and context management. On complex long-horizon tasks (Deep Research and SWE), our folding agent matches or outperforms the ReAct baselines while using an active context 10$\times$ smaller and significantly outperforms models that rely on summarization-based context management.
Similar Papers
AgentFold: Long-Horizon Web Agents with Proactive Context Management
Computation and Language
Helps AI remember more to do complex tasks.
Evaluating Long-Context Reasoning in LLM-Based WebAgents
Machine Learning (CS)
Helps AI remember long conversations to do tasks.
Adaptive Context Length Optimization with Low-Frequency Truncation for Multi-Agent Reinforcement Learning
Machine Learning (CS)
Helps AI teams learn tasks faster and better.