A Diffusion-Refined Planner with Reinforcement Learning Priors for Confined-Space Parking
By: Mingyang Jiang , Yueyuan Li , Jiaru Zhang and more
Potential Business Impact:
Helps cars park perfectly in tight spots.
The growing demand for parking has increased the need for automated parking planning methods that can operate reliably in confined spaces. In restricted and complex environments, high-precision maneuvers are required to achieve a high success rate in planning, yet existing approaches often rely on explicit action modeling, which faces challenges when accurately modeling the optimal action distribution. In this paper, we propose DRIP, a diffusion-refined planner anchored in reinforcement learning (RL) prior action distribution, in which an RL-pretrained policy provides prior action distributions to regularize the diffusion training process. During the inference phase the denoising process refines these coarse priors into more precise action distributions. By steering the denoising trajectory through the reinforcement learning prior distribution during training, the diffusion model inherits a well-informed initialization, resulting in more accurate action modeling, a higher planning success rate, and reduced inference steps. We evaluate our approach across parking scenarios with varying degrees of spatial constraints. Experimental results demonstrate that our method significantly improves planning performance in confined-space parking environments while maintaining strong generalization in common scenarios.
Similar Papers
Automated Parking Trajectory Generation Using Deep Reinforcement Learning
Robotics
Teaches cars to park themselves perfectly.
Dino-Diffusion Modular Designs Bridge the Cross-Domain Gap in Autonomous Parking
Robotics
Helps self-driving cars park anywhere, anytime.
RA-DP: Rapid Adaptive Diffusion Policy for Training-Free High-frequency Robotics Replanning
Robotics
Robots quickly learn new tasks in changing places.