Bootstrapping Reinforcement Learning with Sub-optimal Policies for Autonomous Driving
By: Zhihao Zhang , Chengyang Peng , Ekim Yurtsever and more
Potential Business Impact:
Teaches self-driving cars to learn faster.
Automated vehicle control using reinforcement learning (RL) has attracted significant attention due to its potential to learn driving policies through environment interaction. However, RL agents often face training challenges in sample efficiency and effective exploration, making it difficult to discover an optimal driving strategy. To address these issues, we propose guiding the RL driving agent with a demonstration policy that need not be a highly optimized or expert-level controller. Specifically, we integrate a rule-based lane change controller with the Soft Actor Critic (SAC) algorithm to enhance exploration and learning efficiency. Our approach demonstrates improved driving performance and can be extended to other driving scenarios that can similarly benefit from demonstration-based guidance.
Similar Papers
From Learning to Mastery: Achieving Safe and Efficient Real-World Autonomous Driving with Human-In-The-Loop Reinforcement Learning
Machine Learning (CS)
Teaches self-driving cars to learn safely from humans.
A Comprehensive Review of Reinforcement Learning for Autonomous Driving in the CARLA Simulator
Robotics
Helps self-driving cars learn to drive better.
Automated Parking Trajectory Generation Using Deep Reinforcement Learning
Robotics
Teaches cars to park themselves perfectly.