Consensus-based Decentralized Multi-agent Reinforcement Learning for Random Access Network Optimization
By: Myeung Suk Oh , Zhiyao Zhang , FNU Hairi and more
Potential Business Impact:
Helps many devices share internet without crashing.
With wireless devices increasingly forming a unified smart network for seamless, user-friendly operations, random access (RA) medium access control (MAC) design is considered a key solution for handling unpredictable data traffic from multiple terminals. However, it remains challenging to design an effective RA-based MAC protocol to minimize collisions and ensure transmission fairness across the devices. While existing multi-agent reinforcement learning (MARL) approaches with centralized training and decentralized execution (CTDE) have been proposed to optimize RA performance, their reliance on centralized training and the significant overhead required for information collection can make real-world applications unrealistic. In this work, we adopt a fully decentralized MARL architecture, where policy learning does not rely on centralized tasks but leverages consensus-based information exchanges across devices. We design our MARL algorithm over an actor-critic (AC) network and propose exchanging only local rewards to minimize communication overhead. Furthermore, we provide a theoretical proof of global convergence for our approach. Numerical experiments show that our proposed MARL algorithm can significantly improve RA network performance compared to other baselines.
Similar Papers
Multi-Agent Reinforcement Learning for Task Offloading in Wireless Edge Networks
Machine Learning (CS)
Helps robots share resources without talking much.
Goal-Oriented Multi-Agent Reinforcement Learning for Decentralized Agent Teams
Multiagent Systems
Helps self-driving vehicles work together better.
Multi-Agent Reinforcement Learning for Resources Allocation Optimization: A Survey
Multiagent Systems
Helps computers share resources fairly and smartly.