RoVerFly: Robust and Versatile Learning-based Control of Quadrotor Across Payload Configurations
By: Mintae Kim, Jiaze Cai, Koushil Sreenath
Potential Business Impact:
Drones fly better, even with swinging loads.
Designing robust controllers for precise, arbitrary trajectory tracking with quadrotors is challenging due to nonlinear dynamics and underactuation, and becomes harder with flexible cable-suspended payloads that introduce extra degrees of freedom and hybridness. Classical model-based methods offer stability guarantees but require extensive tuning and often do not adapt when the configuration changes, such as when a payload is added or removed, or when the payload mass or cable length varies. We present RoVerFly, a unified learning-based control framework in which a reinforcement learning (RL) policy serves as a robust and versatile tracking controller for standard quadrotors and for cable-suspended payload systems across a range of configurations. Trained with task and domain randomization, the controller is resilient to disturbances and varying dynamics. It achieves strong zero-shot generalization across payload settings, including no payload as well as varying mass and cable length, without controller switching or re-tuning, while retaining the interpretability and structure of a feedback tracking controller. Code and supplementary materials are available at https://github.com/mintaeshkim/roverfly
Similar Papers
FLARE: Agile Flights for Quadrotor Cable-Suspended Payload System via Reinforcement Learning
Robotics
Drones fly faster and safer through obstacles.
Fault Tolerant Control of a Quadcopter using Reinforcement Learning
Robotics
Keeps drones flying even if a propeller breaks.
Learning Robust Agile Flight Control with Stability Guarantees
Robotics
Lets drones fly faster and safer.