Dynamic one-time delivery of critical data by small and sparse UAV swarms: a model problem for MARL scaling studies
By: Mika Persson , Jonas Lidman , Jacob Ljungberg and more
Potential Business Impact:
Drones learn to deliver packages without crashing.
This work presents a conceptual study on the application of Multi-Agent Reinforcement Learning (MARL) for decentralized control of unmanned aerial vehicles to relay a critical data package to a known position. For this purpose, a family of deterministic games is introduced, designed for scaling studies for MARL. A robust baseline policy is proposed, which is based on restricting agent motion envelopes and applying Dijkstra's algorithm. Experimental results show that two off-the-shelf MARL algorithms perform competitively with the baseline for a small number of agents, but scalability issues arise as the number of agents increase.
Similar Papers
Multi-Agent Deep Reinforcement Learning for Collaborative UAV Relay Networks under Jamming Atatcks
Networking and Internet Architecture
Drones learn to talk without crashing or jamming.
Towards Robust Multi-UAV Collaboration: MARL with Noise-Resilient Communication and Attention Mechanisms
Multiagent Systems
Drones work together better in noisy places.
Scalable UAV Multi-Hop Networking via Multi-Agent Reinforcement Learning with Large Language Models
Multiagent Systems
Drones learn to talk to each other better after disasters.