Learning Closed-Loop Parametric Nash Equilibria of Multi-Agent Collaborative Field Coverage
By: Jushan Chen, Santiago Paternain
Potential Business Impact:
Teaches robots to cover areas much faster.
Multi-agent reinforcement learning is a challenging and active field of research due to the inherent nonstationary property and coupling between agents. A popular approach to modeling the multi-agent interactions underlying the multi-agent RL problem is the Markov Game. There is a special type of Markov Game, termed Markov Potential Game, which allows us to reduce the Markov Game to a single-objective optimal control problem where the objective function is a potential function. In this work, we prove that a multi-agent collaborative field coverage problem, which is found in many engineering applications, can be formulated as a Markov Potential Game, and we can learn a parameterized closed-loop Nash Equilibrium by solving an equivalent single-objective optimal control problem. As a result, our algorithm is 10x faster during training compared to a game-theoretic baseline and converges faster during policy execution.
Similar Papers
From Individual Learning to Market Equilibrium: Correcting Structural and Parametric Biases in RL Simulations of Economic Models
General Economics
Teaches computers to make fair economic choices.
Game Theory and Multi-Agent Reinforcement Learning for Zonal Ancillary Markets
Multiagent Systems
Makes electricity grids work better and cheaper.
Heuristics for Combinatorial Optimization via Value-based Reinforcement Learning: A Unified Framework and Analysis
Machine Learning (Stat)
Helps computers solve hard puzzles faster and better.