Score: 0

Hierarchical Reinforcement Learning with Low-Level MPC for Multi-Agent Control

Published: September 19, 2025 | arXiv ID: 2509.15799v1

By: Max Studt, Georg Schildbach

Potential Business Impact:

Helps robots learn to move safely together.

Business Areas:
Multi-level Marketing Sales and Marketing

Achieving safe and coordinated behavior in dynamic, constraint-rich environments remains a major challenge for learning-based control. Pure end-to-end learning often suffers from poor sample efficiency and limited reliability, while model-based methods depend on predefined references and struggle to generalize. We propose a hierarchical framework that combines tactical decision-making via reinforcement learning (RL) with low-level execution through Model Predictive Control (MPC). For the case of multi-agent systems this means that high-level policies select abstract targets from structured regions of interest (ROIs), while MPC ensures dynamically feasible and safe motion. Tested on a predator-prey benchmark, our approach outperforms end-to-end and shielding-based RL baselines in terms of reward, safety, and consistency, underscoring the benefits of combining structured learning with model-based control.

Page Count
7 pages

Category
Electrical Engineering and Systems Science:
Systems and Control