Engineered over Emergent Communication in MARL for Scalable and Sample-Efficient Cooperative Task Allocation in a Partially Observable Grid
By: Brennen A. Hill, Mant Koh En Wei, Thangavel Jishnuanandh
Potential Business Impact:
Helps robots work together better by planning ahead.
We compare the efficacy of learned versus engineered communication strategies in a cooperative multi-agent reinforcement learning (MARL) environment. For the learned approach, we introduce Learned Direct Communication (LDC), where agents generate messages and actions concurrently via a neural network. Our engineered approach, Intention Communication, employs an Imagined Trajectory Generation Module (ITGM) and a Message Generation Network (MGN) to formulate messages based on predicted future states. Both strategies are evaluated on their success rates in cooperative tasks under fully and partially observable conditions. Our findings indicate that while emergent communication is viable, the engineered approach demonstrates superior performance and scalability, particularly as environmental complexity increases.
Similar Papers
Learning what to say and how precisely: Efficient Communication via Differentiable Discrete Communication Learning
Multiagent Systems
Agents learn to send smarter, smaller messages.
Learning Efficient Communication Protocols for Multi-Agent Reinforcement Learning
Multiagent Systems
Teaches robots to talk smarter, work better together.
Language-Guided Multi-Agent Learning in Simulations: A Unified Framework and Evaluation
Artificial Intelligence
Helps AI teams work together better in games.