Learning Natural and Robust Hexapod Locomotion over Complex Terrains via Motion Priors based on Deep Reinforcement Learning
By: Xin Liu , Jinze Wu , Yinghui Li and more
Potential Business Impact:
Robots walk better on rough ground.
Multi-legged robots offer enhanced stability to navigate complex terrains with their multiple legs interacting with the environment. However, how to effectively coordinate the multiple legs in a larger action exploration space to generate natural and robust movements is a key issue. In this paper, we introduce a motion prior-based approach, successfully applying deep reinforcement learning algorithms to a real hexapod robot. We generate a dataset of optimized motion priors, and train an adversarial discriminator based on the priors to guide the hexapod robot to learn natural gaits. The learned policy is then successfully transferred to a real hexapod robot, and demonstrate natural gait patterns and remarkable robustness without visual information in complex terrains. This is the first time that a reinforcement learning controller has been used to achieve complex terrain walking on a real hexapod robot.
Similar Papers
MoRE: Mixture of Residual Experts for Humanoid Lifelike Gaits Learning on Complex Terrains
Robotics
Robots walk like people on any ground.
Robust Humanoid Walking on Compliant and Uneven Terrain with Deep Reinforcement Learning
Robotics
Robots learn to walk on bumpy, soft ground.
Teacher Motion Priors: Enhancing Robot Locomotion over Challenging Terrain
Robotics
Robot learns to walk on tricky ground better.