Real-Time Gait Adaptation for Quadrupeds using Model Predictive Control and Reinforcement Learning
By: Ganga Nair B, Prakrut Kotecha, Shishir Kolathaya
Potential Business Impact:
Robot dogs walk better and use less energy.
Model-free reinforcement learning (RL) has enabled adaptable and agile quadruped locomotion; however, policies often converge to a single gait, leading to suboptimal performance. Traditionally, Model Predictive Control (MPC) has been extensively used to obtain task-specific optimal policies but lacks the ability to adapt to varying environments. To address these limitations, we propose an optimization framework for real-time gait adaptation in a continuous gait space, combining the Model Predictive Path Integral (MPPI) algorithm with a Dreamer module to produce adaptive and optimal policies for quadruped locomotion. At each time step, MPPI jointly optimizes the actions and gait variables using a learned Dreamer reward that promotes velocity tracking, energy efficiency, stability, and smooth transitions, while penalizing abrupt gait changes. A learned value function is incorporated as terminal reward, extending the formulation to an infinite-horizon planner. We evaluate our framework in simulation on the Unitree Go1, demonstrating an average reduction of up to 36.48\% in energy consumption across varying target speeds, while maintaining accurate tracking and adaptive, task-appropriate gaits.
Similar Papers
Real-Time Gait Adaptation for Quadrupeds using Model Predictive Control and Reinforcement Learning
Robotics
Robots walk better and use less energy.
Reference-Free Sampling-Based Model Predictive Control
Robotics
Robots learn to walk, jump, and balance themselves.
Adaptive Legged Locomotion via Online Learning for Model Predictive Control
Robotics
Robots learn to walk on tricky ground.