Walk the PLANC: Physics-Guided RL for Agile Humanoid Locomotion on Constrained Footholds
By: Min Dai , William D. Compton , Junheng Li and more
Bipedal humanoid robots must precisely coordinate balance, timing, and contact decisions when locomoting on constrained footholds such as stepping stones, beams, and planks -- even minor errors can lead to catastrophic failure. Classical optimization and control pipelines handle these constraints well but depend on highly accurate mathematical representations of terrain geometry, making them prone to error when perception is noisy or incomplete. Meanwhile, reinforcement learning has shown strong resilience to disturbances and modeling errors, yet end-to-end policies rarely discover the precise foothold placement and step sequencing required for discontinuous terrain. These contrasting limitations motivate approaches that guide learning with physics-based structure rather than relying purely on reward shaping. In this work, we introduce a locomotion framework in which a reduced-order stepping planner supplies dynamically consistent motion targets that steer the RL training process via Control Lyapunov Function (CLF) rewards. This combination of structured footstep planning and data-driven adaptation produces accurate, agile, and hardware-validated stepping-stone locomotion on a humanoid robot, substantially improving reliability compared to conventional model-free reinforcement-learning baselines.
Similar Papers
Chasing Stability: Humanoid Running via Control Lyapunov Function Guided Reinforcement Learning
Robotics
Robot learns to run and stay balanced.
CLF-RL: Control Lyapunov Function Guided Reinforcement Learning
Robotics
Helps robots walk better without falling.
Learning a Vision-Based Footstep Planner for Hierarchical Walking Control
Robotics
Robots walk better on bumpy ground using sight.