An Initial Study of Bird's-Eye View Generation for Autonomous Vehicles using Cross-View Transformers
By: Felipe Carlos dos Santos, Eric Aislan Antonelo, Gustavo Claudio Karl Couto
Potential Business Impact:
Helps self-driving cars see roads from above.
Bird's-Eye View (BEV) maps provide a structured, top-down abstraction that is crucial for autonomous-driving perception. In this work, we employ Cross-View Transformers (CVT) for learning to map camera images to three BEV's channels - road, lane markings, and planned trajectory - using a realistic simulator for urban driving. Our study examines generalization to unseen towns, the effect of different camera layouts, and two loss formulations (focal and L1). Using training data from only a town, a four-camera CVT trained with the L1 loss delivers the most robust test performance, evaluated in a new town. Overall, our results underscore CVT's promise for mapping camera inputs to reasonably accurate BEV maps.
Similar Papers
Progressive Bird's Eye View Perception for Safety-Critical Autonomous Driving: A Comprehensive Survey
Robotics
Makes self-driving cars safer in bad weather.
BEVTraj: Map-Free End-to-End Trajectory Prediction in Bird's-Eye View with Deformable Attention and Sparse Goal Proposals
CV and Pattern Recognition
Helps self-driving cars see without maps.
BEVCon: Advancing Bird's Eye View Perception with Contrastive Learning
CV and Pattern Recognition
Helps self-driving cars see better from above.