Local-Canonicalization Equivariant Graph Neural Networks for Sample-Efficient and Generalizable Swarm Robot Control
By: Keqin Wang , Tao Zhong , David Chang and more
Potential Business Impact:
Helps robot teams work together better, even when things change.
Multi-agent reinforcement learning (MARL) has emerged as a powerful paradigm for coordinating swarms of agents in complex decision-making, yet major challenges remain. In competitive settings such as pursuer-evader tasks, simultaneous adaptation can destabilize training; non-kinetic countermeasures often fail under adverse conditions; and policies trained in one configuration rarely generalize to environments with a different number of agents. To address these issues, we propose the Local-Canonicalization Equivariant Graph Neural Networks (LEGO) framework, which integrates seamlessly with popular MARL algorithms such as MAPPO. LEGO employs graph neural networks to capture permutation equivariance and generalization to different agent numbers, canonicalization to enforce E(n)-equivariance, and heterogeneous representations to encode role-specific inductive biases. Experiments on cooperative and competitive swarm benchmarks show that LEGO outperforms strong baselines and improves generalization. In real-world experiments, LEGO demonstrates robustness to varying team sizes and agent failure.
Similar Papers
Q-MARL: A quantum-inspired algorithm using neural message passing for large-scale multi-agent reinforcement learning
Machine Learning (CS)
Lets many robots learn to work together.
Eq.Bot: Enhance Robotic Manipulation Learning via Group Equivariant Canonicalization
Robotics
Robots learn to move objects more accurately.
Bayesian Ego-graph inference for Networked Multi-Agent Reinforcement Learning
Multiagent Systems
Helps many robots learn to work together better.