On Architectures for Combining Reinforcement Learning and Model Predictive Control with Runtime Improvements
By: Xiaolong Jia, Nikhil Bajaj
Potential Business Impact:
Makes robots learn faster and control better.
Model Predictive Control (MPC) faces computational demands and performance degradation from model inaccuracies. We propose two architectures combining Neural Network-approximated MPC (NNMPC) with Reinforcement Learning (RL). The first, Warm Start RL, initializes the RL actor with pre-trained NNMPC weights. The second, RLMPC, uses RL to generate corrective residuals for NNMPC outputs. We introduce a downsampling method reducing NNMPC input dimensions while maintaining performance. Evaluated on a rotary inverted pendulum, both architectures demonstrate runtime reductions exceeding 99% compared to traditional MPC while improving tracking performance under model uncertainties, with RL+MPC achieving 11-40% cost reduction depending on reference amplitude.
Similar Papers
Residual MPC: Blending Reinforcement Learning with GPU-Parallelized Model Predictive Control
Robotics
Robots walk better by combining two smart control methods.
Model Predictive Control-Guided Reinforcement Learning for Implicit Balancing
Systems and Control
Smarter batteries earn more money for power grids.
Hybrid Reinforcement Learning and Model Predictive Control for Adaptive Control of Hydrogen-Diesel Dual-Fuel Combustion
Machine Learning (CS)
Makes engines run better and safer, even when they change.