Reasoning-VLA: A Fast and General Vision-Language-Action Reasoning Model for Autonomous Driving
By: Dapeng Zhang , Zhenlong Yuan , Zhangquan Chen and more
Potential Business Impact:
Helps self-driving cars drive smarter and faster.
Vision-Language-Action (VLA) models have recently shown strong decision-making capabilities in autonomous driving. However, existing VLAs often struggle with achieving efficient inference and generalizing to novel autonomous vehicle configurations and driving scenarios. In this paper, we propose Reasoning-VLA, a general and fast action-generation VLA framework. The proposed model employs a set of learnable action queries, initialized via Gaussian sampling from ground-truth trajectories within the training corpus. These learnable queries interact with reasoning-enhanced vision-language features to generate continuous action trajectories in parallel. To promote robust generalization, we consolidate eight publicly available autonomous driving datasets into a standardized, Chain-of-Thought reasoning-based, and easy-to-use data format for model training. Leveraging both supervised learning and reinforcement learning fine-tuning, extensive empirical evaluations across multiple benchmarks demonstrate that Reasoning-VLA achieves state-of-the-art performance, superior generalization capability, and the excellent inference speed reported to date.
Similar Papers
AutoVLA: A Vision-Language-Action Model for End-to-End Autonomous Driving with Adaptive Reasoning and Reinforcement Fine-Tuning
CV and Pattern Recognition
Helps self-driving cars plan safer, faster trips.
VLA-R1: Enhancing Reasoning in Vision-Language-Action Models
CV and Pattern Recognition
Teaches robots to think and do tasks.
DualVLA: Building a Generalizable Embodied Agent via Partial Decoupling of Reasoning and Action
CV and Pattern Recognition
Teaches robots to act and think better.