Multi-Agent Reinforcement Learning with Communication-Constrained Priors
By: Guang Yang , Tianpei Yang , Jingwen Qiao and more
Potential Business Impact:
Helps robots learn to work together even with bad signals.
Communication is one of the effective means to improve the learning of cooperative policy in multi-agent systems. However, in most real-world scenarios, lossy communication is a prevalent issue. Existing multi-agent reinforcement learning with communication, due to their limited scalability and robustness, struggles to apply to complex and dynamic real-world environments. To address these challenges, we propose a generalized communication-constrained model to uniformly characterize communication conditions across different scenarios. Based on this, we utilize it as a learning prior to distinguish between lossy and lossless messages for specific scenarios. Additionally, we decouple the impact of lossy and lossless messages on distributed decision-making, drawing on a dual mutual information estimatior, and introduce a communication-constrained multi-agent reinforcement learning framework, quantifying the impact of communication messages into the global reward. Finally, we validate the effectiveness of our approach across several communication-constrained benchmarks.
Similar Papers
Toward Goal-Oriented Communication in Multi-Agent Systems: An overview
Multiagent Systems
Helps robots work together smarter with less talking.
Learning Efficient Communication Protocols for Multi-Agent Reinforcement Learning
Multiagent Systems
Teaches robots to talk smarter, work better together.
Information-Theoretic Limits of Integrated Sensing and Communication with Finite Learning Capacity
Information Theory
AI helps devices share data and sense surroundings.