Discovering Coordinated Joint Options via Inter-Agent Relative Dynamics
By: Raul D. Steleac, Mohan Sridharan, David Abel
Temporally extended actions improve the ability to explore and plan in single-agent settings. In multi-agent settings, the exponential growth of the joint state space with the number of agents makes coordinated behaviours even more valuable. Yet, this same exponential growth renders the design of multi-agent options particularly challenging. Existing multi-agent option discovery methods often sacrifice coordination by producing loosely coupled or fully independent behaviours. Toward addressing these limitations, we describe a novel approach for multi-agent option discovery. Specifically, we propose a joint-state abstraction that compresses the state space while preserving the information necessary to discover strongly coordinated behaviours. Our approach builds on the inductive bias that synchronisation over agent states provides a natural foundation for coordination in the absence of explicit objectives. We first approximate a fictitious state of maximal alignment with the team, the \textit{Fermat} state, and use it to define a measure of \textit{spreadness}, capturing team-level misalignment on each individual state dimension. Building on this representation, we then employ a neural graph Laplacian estimator to derive options that capture state synchronisation patterns between agents. We evaluate the resulting options across multiple scenarios in two multi-agent domains, showing that they yield stronger downstream coordination capabilities compared to alternative option discovery methods.
Similar Papers
Multi-Action Self-Improvement for Neural Combinatorial Optimization
Machine Learning (CS)
Teaches computers to solve complex problems faster.
An Efficient Approach for Cooperative Multi-Agent Learning Problems
Artificial Intelligence
Teaches robots to work together better.
Situationally-Aware Dynamics Learning
Robotics
Robots learn to handle unexpected problems safely.