SafeVLA: Towards Safety Alignment of Vision-Language-Action Model via Constrained Learning
By: Borong Zhang , Yuhao Zhang , Jiaming Ji and more
Potential Business Impact:
Keeps robots safe while they learn tasks.
Vision-language-action models (VLAs) show potential as generalist robot policies. However, these models pose extreme safety challenges during real-world deployment, including the risk of harm to the environment, the robot itself, and humans. How can safety constraints be explicitly integrated into VLAs? We address this by exploring an integrated safety approach (ISA), systematically modeling safety requirements, then actively eliciting diverse unsafe behaviors, effectively constraining VLA policies via safe reinforcement learning, and rigorously assuring their safety through targeted evaluations. Leveraging the constrained Markov decision process (CMDP) paradigm, ISA optimizes VLAs from a min-max perspective against elicited safety risks. Thus, policies aligned through this comprehensive approach achieve the following key features: (I) effective safety-performance trade-offs, this exploration yields an 83.58% safety improvement compared to the current state-of-the-art method, while also maintaining task performance (+3.85%). (II) strong safety assurance, with the ability to mitigate long-tail risks and handle extreme failure scenarios. (III) robust generalization of learned safety behaviors to various out-of-distribution perturbations. Our data, models and newly proposed benchmark environment are available at https://pku-safevla.github.io.
Similar Papers
VLSA: Vision-Language-Action Models with Plug-and-Play Safety Constraint Layer
Robotics
Keeps robots safe while doing tasks.
Pure Vision Language Action (VLA) Models: A Comprehensive Survey
Robotics
Robots learn to see, talk, and do tasks.
A Vision-Language-Action-Critic Model for Robotic Real-World Reinforcement Learning
Robotics
Helps robots learn tasks faster and better.