Reactive Aerobatic Flight via Reinforcement Learning
By: Zhichao Han , Xijie Huang , Zhuxiu Xu and more
Potential Business Impact:
Drones learn to do amazing flips and loops.
Quadrotors have demonstrated remarkable versatility, yet their full aerobatic potential remains largely untapped due to inherent underactuation and the complexity of aggressive maneuvers. Traditional approaches, separating trajectory optimization and tracking control, suffer from tracking inaccuracies, computational latency, and sensitivity to initial conditions, limiting their effectiveness in dynamic, high-agility scenarios. Inspired by recent breakthroughs in data-driven methods, we propose a reinforcement learning-based framework that directly maps drone states and aerobatic intentions to control commands, eliminating modular separation to enable quadrotors to perform end-to-end policy optimization for extreme aerobatic maneuvers. To ensure efficient and stable training, we introduce an automated curriculum learning strategy that dynamically adjusts aerobatic task difficulty. Enabled by domain randomization for robust zero-shot sim-to-real transfer, our approach is validated in demanding real-world experiments, including the first demonstration of a drone autonomously performing continuous inverted flight while reactively navigating a moving gate, showcasing unprecedented agility.
Similar Papers
Dream to Fly: Model-Based Reinforcement Learning for Vision-Based Drone Flight
Robotics
Drones fly race tracks using only their eyes.
Dynamics-Invariant Quadrotor Control using Scale-Aware Deep Reinforcement Learning
Systems and Control
Drones fly better even with wind and heavy loads.
Equivariant Reinforcement Learning Frameworks for Quadrotor Low-Level Control
Robotics
Drones learn to fly better with less practice.