Cooperative UAVs for Remote Data Collection under Limited Communications: An Asynchronous Multiagent Learning Framework
By: Cuong Le, Symeon Chatzinotas, Thang X. Vu
Potential Business Impact:
Drones collect data better, using less power.
This paper addresses the joint optimization of trajectories and bandwidth allocation for multiple Unmanned Aerial Vehicles (UAVs) to enhance energy efficiency in the cooperative data collection problem. We focus on an important yet underestimated aspect of the system, where action synchronization across all UAVs is impossible. Since most existing learning-based solutions are not designed to learn in this asynchronous environment, we formulate the trajectory planning problem as a Decentralized Partially Observable Semi-Markov Decision Process and introduce an asynchronous multi-agent learning algorithm to learn UAVs' cooperative policies. Once the UAVs' trajectory policies are learned, the bandwidth allocation can be optimally solved based on local observations at each collection point. Comprehensive empirical results demonstrate the superiority of the proposed method over other learning-based and heuristic baselines in terms of both energy efficiency and mission completion time. Additionally, the learned policies exhibit robustness under varying environmental conditions.
Similar Papers
Heterogeneous Multi-agent Collaboration in UAV-assisted Mobile Crowdsensing Networks
Machine Learning (CS)
Drones collect more data by planning better.
Multi-Agent Deep Reinforcement Learning for Collaborative UAV Relay Networks under Jamming Atatcks
Networking and Internet Architecture
Drones learn to talk without crashing or jamming.
Decentralized Real-Time Planning for Multi-UAV Cooperative Manipulation via Imitation Learning
Robotics
Drones carry heavy things without talking to each other.