Learning Mean-Field Games through Mean-Field Actor-Critic Flow
By: Mo Zhou, Haosheng Zhou, Ruimeng Hu
Potential Business Impact:
Teaches computers to make smart group decisions.
We propose the Mean-Field Actor-Critic (MFAC) flow, a continuous-time learning dynamics for solving mean-field games (MFGs), combining techniques from reinforcement learning and optimal transport. The MFAC framework jointly evolves the control (actor), value function (critic), and distribution components through coupled gradient-based updates governed by partial differential equations (PDEs). A central innovation is the Optimal Transport Geodesic Picard (OTGP) flow, which drives the distribution toward equilibrium along Wasserstein-2 geodesics. We conduct a rigorous convergence analysis using Lyapunov functionals and establish global exponential convergence of the MFAC flow under a suitable timescale. Our results highlight the algorithmic interplay among actor, critic, and distribution components. Numerical experiments illustrate the theoretical findings and demonstrate the effectiveness of the MFAC framework in computing MFG equilibria.
Similar Papers
High-dimensional Mean-Field Games by Particle-based Flow Matching
Machine Learning (Stat)
Solves hard math problems for many moving parts.
Learning in Stackelberg Mean Field Games: A Non-Asymptotic Analysis
Machine Learning (CS)
Helps one player beat many by predicting their moves.
Mean-Field Games with Constraints
Optimization and Control
Helps many players make good choices together.