Quadrotor Navigation using Reinforcement Learning with Privileged Information
By: Jonathan Lee , Abhishek Rathod , Kshitij Goel and more
Potential Business Impact:
Drones fly around big walls without crashing.
This paper presents a reinforcement learning-based quadrotor navigation method that leverages efficient differentiable simulation, novel loss functions, and privileged information to navigate around large obstacles. Prior learning-based methods perform well in scenes that exhibit narrow obstacles, but struggle when the goal location is blocked by large walls or terrain. In contrast, the proposed method utilizes time-of-arrival (ToA) maps as privileged information and a yaw alignment loss to guide the robot around large obstacles. The policy is evaluated in photo-realistic simulation environments containing large obstacles, sharp corners, and dead-ends. Our approach achieves an 86% success rate and outperforms baseline strategies by 34%. We deploy the policy onboard a custom quadrotor in outdoor cluttered environments both during the day and night. The policy is validated across 20 flights, covering 589 meters without collisions at speeds up to 4 m/s.
Similar Papers
Learning Obstacle Avoidance using Double DQN for Quadcopter Navigation
Robotics
Drones learn to fly safely in cities.
LEARN: Learning End-to-End Aerial Resource-Constrained Multi-Robot Navigation
Robotics
Tiny drones fly safely through tight spaces.
Autonomous UAV Flight Navigation in Confined Spaces: A Reinforcement Learning Approach
Robotics
Drones learn to fly safely in dark tunnels.