Self-Supervised Goal-Reaching Results in Multi-Agent Cooperation and Exploration
By: Chirayu Nimonkar , Shlok Shah , Catherine Ji and more
Potential Business Impact:
Robots learn to work together to reach goals.
For groups of autonomous agents to achieve a particular goal, they must engage in coordination and long-horizon reasoning. However, designing reward functions to elicit such behavior is challenging. In this paper, we study how self-supervised goal-reaching techniques can be leveraged to enable agents to cooperate. The key idea is that, rather than have agents maximize some scalar reward, agents aim to maximize the likelihood of visiting a certain goal. This problem setting enables human users to specify tasks via a single goal state rather than implementing a complex reward function. While the feedback signal is quite sparse, we will demonstrate that self-supervised goal-reaching techniques enable agents to learn from such feedback. On MARL benchmarks, our proposed method outperforms alternative approaches that have access to the same sparse reward signal as our method. While our method has no explicit mechanism for exploration, we observe that self-supervised multi-agent goal-reaching leads to emergent cooperation and exploration in settings where alternative approaches never witness a single successful trial.
Similar Papers
Demystifying the Mechanisms Behind Emergent Exploration in Goal-conditioned RL
Machine Learning (CS)
Teaches robots to explore safely without being told.
Goal-Oriented Multi-Agent Reinforcement Learning for Decentralized Agent Teams
Multiagent Systems
Helps self-driving vehicles work together better.
Search Self-play: Pushing the Frontier of Agent Capability without Supervision
Machine Learning (CS)
Teaches AI to learn by playing against itself.