Guided by Guardrails: Control Barrier Functions as Safety Instructors for Robotic Learning
By: Maeva Guerrier , Karthik Soma , Hassan Fouad and more
Potential Business Impact:
Teaches robots to learn safely without crashing.
Safety stands as the primary obstacle preventing the widespread adoption of learning-based robotic systems in our daily lives. While reinforcement learning (RL) shows promise as an effective robot learning paradigm, conventional RL frameworks often model safety by using single scalar negative rewards with immediate episode termination, failing to capture the temporal consequences of unsafe actions (e.g., sustained collision damage). In this work, we introduce a novel approach that simulates these temporal effects by applying continuous negative rewards without episode termination. Our experiments reveal that standard RL methods struggle with this model, as the accumulated negative values in unsafe zones create learning barriers. To address this challenge, we demonstrate how Control Barrier Functions (CBFs), with their proven safety guarantees, effectively help robots avoid catastrophic regions while enhancing learning outcomes. We present three CBF-based approaches, each integrating traditional RL methods with Control Barrier Functions, guiding the agent to learn safe behavior. Our empirical analysis, conducted in both simulated environments and real-world settings using a four-wheel differential drive robot, explores the possibilities of employing these approaches for safe robotic learning.
Similar Papers
CBF-RL: Safety Filtering Reinforcement Learning in Training with Control Barrier Functions
Robotics
Teaches robots to be safe while learning.
CBF-RL: Safety Filtering Reinforcement Learning in Training with Control Barrier Functions
Robotics
Teaches robots to be safe while learning.
Learning Neural Control Barrier Functions from Expert Demonstrations using Inverse Constraint Learning
Artificial Intelligence
Teaches robots to avoid danger using examples.