Learning Decentralized Routing Policies via Graph Attention-based Multi-Agent Reinforcement Learning in Lunar Delay-Tolerant Networks
By: Federico Lozano-Cuadra , Beatriz Soret , Marc Sanchez Net and more
Potential Business Impact:
Robots on the Moon share data despite bad connections.
We present a fully decentralized routing framework for multi-robot exploration missions operating under the constraints of a Lunar Delay-Tolerant Network (LDTN). In this setting, autonomous rovers must relay collected data to a lander under intermittent connectivity and unknown mobility patterns. We formulate the problem as a Partially Observable Markov Decision Problem (POMDP) and propose a Graph Attention-based Multi-Agent Reinforcement Learning (GAT-MARL) policy that performs Centralized Training, Decentralized Execution (CTDE). Our method relies only on local observations and does not require global topology updates or packet replication, unlike classical approaches such as shortest path and controlled flooding-based algorithms. Through Monte Carlo simulations in randomized exploration environments, GAT-MARL provides higher delivery rates, no duplications, and fewer packet losses, and is able to leverage short-term mobility forecasts; offering a scalable solution for future space robotic systems for planetary exploration, as demonstrated by successful generalization to larger rover teams.
Similar Papers
Goal-Oriented Multi-Agent Reinforcement Learning for Decentralized Agent Teams
Multiagent Systems
Helps self-driving vehicles work together better.
Multi-Agent Path Finding via Offline RL and LLM Collaboration
Multiagent Systems
Robots learn to move together faster, smarter.
Structured Cooperative Multi-Agent Reinforcement Learning: a Bayesian Network Perspective
Multiagent Systems
Helps many robots learn to work together better.