Multi-Agent Collaboration via Evolving Orchestration
By: Yufan Dang , Chen Qian , Xueheng Luo and more
Potential Business Impact:
Lets AI teams work together smarter, faster.
Large language models (LLMs) have achieved remarkable results across diverse downstream tasks, but their monolithic nature restricts scalability and efficiency in complex problem-solving. While recent research explores multi-agent collaboration among LLMs, most approaches rely on static organizational structures that struggle to adapt as task complexity and agent numbers grow, resulting in coordination overhead and inefficiencies. To this end, we propose a puppeteer-style paradigm for LLM-based multi-agent collaboration, where a centralized orchestrator ("puppeteer") dynamically directs agents ("puppets") in response to evolving task states. This orchestrator is trained via reinforcement learning to adaptively sequence and prioritize agents, enabling flexible and evolvable collective reasoning. Experiments on closed- and open-domain scenarios show that this method achieves superior performance with reduced computational costs. Analyses further reveal that the key improvements consistently stem from the emergence of more compact, cyclic reasoning structures under the orchestrator's evolution.
Similar Papers
Multi-Agent Systems for Robotic Autonomy with LLMs
Robotics
Builds robots that can do jobs by themselves.
Multi-Agent Collaboration Mechanisms: A Survey of LLMs
Artificial Intelligence
Lets AI groups work together to solve hard problems.
Multi-Agent Language Models: Advancing Cooperation, Coordination, and Adaptation
Computation and Language
Helps AI understand and work with people.