Robust RL Control for Bipedal Locomotion with Closed Kinematic Chains
By: Egor Maslennikov , Eduard Zaliaev , Nikita Dudorov and more
Potential Business Impact:
Robots walk better on bumpy ground.
Developing robust locomotion controllers for bipedal robots with closed kinematic chains presents unique challenges, particularly since most reinforcement learning (RL) approaches simplify these parallel mechanisms into serial models during training. We demonstrate that this simplification significantly impairs sim-to-real transfer by failing to capture essential aspects such as joint coupling, friction dynamics, and motor-space control characteristics. In this work, we present an RL framework that explicitly incorporates closed-chain dynamics and validate it on our custom-built robot TopA. Our approach enhances policy robustness through symmetry-aware loss functions, adversarial training, and targeted network regularization. Experimental results demonstrate that our integrated approach achieves stable locomotion across diverse terrains, significantly outperforming methods based on simplified kinematic models.
Similar Papers
Optimal Control of Walkers with Parallel Actuation
Robotics
Helps robots with many legs move better.
Efficient Learning of A Unified Policy For Whole-body Manipulation and Locomotion Skills
Robotics
Robots learn to move and grab better.
Robust Humanoid Walking on Compliant and Uneven Terrain with Deep Reinforcement Learning
Robotics
Robots learn to walk on bumpy, soft ground.