Evaluating Robustness of Deep Reinforcement Learning for Autonomous Surface Vehicle Control in Field Tests
By: Luis F. W. Batista , Stéphanie Aravecchia , Seth Hutchinson and more
Potential Business Impact:
Robot boats collect trash even when pushed.
Despite significant advancements in Deep Reinforcement Learning (DRL) for Autonomous Surface Vehicles (ASVs), their robustness in real-world conditions, particularly under external disturbances, remains insufficiently explored. In this paper, we evaluate the resilience of a DRL-based agent designed to capture floating waste under various perturbations. We train the agent using domain randomization and evaluate its performance in real-world field tests, assessing its ability to handle unexpected disturbances such as asymmetric drag and an off-center payload. We assess the agent's performance under these perturbations in both simulation and real-world experiments, quantifying performance degradation and benchmarking it against an MPC baseline. Results indicate that the DRL agent performs reliably despite significant disturbances. Along with the open-source release of our implementation, we provide insights into effective training strategies, real-world challenges, and practical considerations for deploying DRLbased ASV controllers.
Similar Papers
Stability Analysis of Deep Reinforcement Learning for Multi-Agent Inspection in a Terrestrial Testbed
Robotics
Lets robots in space work together without mistakes.
Depth-Constrained ASV Navigation with Deep RL and Limited Sensing
Robotics
Helps boats safely steer in shallow water.
Control-Optimized Deep Reinforcement Learning for Artificially Intelligent Autonomous Systems
Robotics
AI learns to fix its own mistakes.