Safe Heterogeneous Multi-Agent RL with Communication Regularization for Coordinated Target Acquisition
By: Gabriele Calzolari , Vidya Sumathy , Christoforos Kanellakis and more
This paper introduces a decentralized multi-agent reinforcement learning framework enabling structurally heterogeneous teams of agents to jointly discover and acquire randomly located targets in environments characterized by partial observability, communication constraints, and dynamic interactions. Each agent's policy is trained with the Multi-Agent Proximal Policy Optimization algorithm and employs a Graph Attention Network encoder that integrates simulated range-sensing data with communication embeddings exchanged among neighboring agents, enabling context-aware decision-making from both local sensing and relational information. In particular, this work introduces a unified framework that integrates graph-based communication and trajectory-aware safety through safety filters. The architecture is supported by a structured reward formulation designed to encourage effective target discovery and acquisition, collision avoidance, and de-correlation between the agents' communication vectors by promoting informational orthogonality. The effectiveness of the proposed reward function is demonstrated through a comprehensive ablation study. Moreover, simulation results demonstrate safe and stable task execution, confirming the framework's effectiveness.
Similar Papers
Platform-Agnostic Reinforcement Learning Framework for Safe Exploration of Cluttered Environments with Graph Attention
Robotics
Helps robots explore dangerous places safely.
Goal-Oriented Multi-Agent Reinforcement Learning for Decentralized Agent Teams
Multiagent Systems
Helps self-driving vehicles work together better.
Learning to Communicate in Multi-Agent Reinforcement Learning for Autonomous Cyber Defence
Multiagent Systems
Teaches AI to talk and stop cyberattacks.