Scaling Rough Terrain Locomotion with Automatic Curriculum Reinforcement Learning
By: Ziming Li, Chenhao Li, Marco Hutter
Potential Business Impact:
Robot learns to walk anywhere, fast.
Curriculum learning has demonstrated substantial effectiveness in robot learning. However, it still faces limitations when scaling to complex, wide-ranging task spaces. Such task spaces often lack a well-defined difficulty structure, making the difficulty ordering required by previous methods challenging to define. We propose a Learning Progress-based Automatic Curriculum Reinforcement Learning (LP-ACRL) framework, which estimates the agent's learning progress online and adaptively adjusts the task-sampling distribution, thereby enabling automatic curriculum generation without prior knowledge of the difficulty distribution over the task space. Policies trained with LP-ACRL enable the ANYmal D quadruped to achieve and maintain stable, high-speed locomotion at 2.5 m/s linear velocity and 3.0 rad/s angular velocity across diverse terrains, including stairs, slopes, gravel, and low-friction flat surfaces--whereas previous methods have generally been limited to high speeds on flat terrain or low speeds on complex terrain. Experimental results demonstrate that LP-ACRL exhibits strong scalability and real-world applicability, providing a robust baseline for future research on curriculum generation in complex, wide-ranging robotic learning task spaces.
Similar Papers
GACL: Grounded Adaptive Curriculum Learning with Active Task and Performance Monitoring
Robotics
Teaches robots new skills faster and better.
Automatic Curriculum Learning for Driving Scenarios: Towards Robust and Efficient Reinforcement Learning
Robotics
Teaches self-driving cars to learn better.
AdaCuRL: Adaptive Curriculum Reinforcement Learning with Invalid Sample Mitigation and Historical Revisiting
Machine Learning (CS)
Teaches AI to learn harder tasks better.