LIPM-Guided Reinforcement Learning for Stable and Perceptive Locomotion in Bipedal Robots
By: Haokai Su , Haoxiang Luo , Shunpeng Yang and more
Potential Business Impact:
Robots walk steadily on bumpy ground.
Achieving stable and robust perceptive locomotion for bipedal robots in unstructured outdoor environments remains a critical challenge due to complex terrain geometry and susceptibility to external disturbances. In this work, we propose a novel reward design inspired by the Linear Inverted Pendulum Model (LIPM) to enable perceptive and stable locomotion in the wild. The LIPM provides theoretical guidance for dynamic balance by regulating the center of mass (CoM) height and the torso orientation. These are key factors for terrain-aware locomotion, as they help ensure a stable viewpoint for the robot's camera. Building on this insight, we design a reward function that promotes balance and dynamic stability while encouraging accurate CoM trajectory tracking. To adaptively trade off between velocity tracking and stability, we leverage the Reward Fusion Module (RFM) approach that prioritizes stability when needed. A double-critic architecture is adopted to separately evaluate stability and locomotion objectives, improving training efficiency and robustness. We validate our approach through extensive experiments on a bipedal robot in both simulation and real-world outdoor environments. The results demonstrate superior terrain adaptability, disturbance rejection, and consistent performance across a wide range of speeds and perceptual conditions.
Similar Papers
Bipedal Robust Walking on Uneven Footholds: Piecewise Slope LIPM with Discrete Model Predictive Control
Robotics
Helps robots walk on bumpy ground.
CLF-RL: Control Lyapunov Function Guided Reinforcement Learning
Robotics
Helps robots walk better without falling.
RL-augmented Adaptive Model Predictive Control for Bipedal Locomotion over Challenging Terrain
Robotics
Helps robots walk safely on slippery, uneven ground.