ViSE: A Systematic Approach to Vision-Only Street-View Extrapolation
By: Kaiyuan Tan , Yingying Shen , Haiyang Sun and more
Potential Business Impact:
Helps self-driving cars see around corners.
Realistic view extrapolation is critical for closed-loop simulation in autonomous driving, yet it remains a significant challenge for current Novel View Synthesis (NVS) methods, which often produce distorted and inconsistent images beyond the original trajectory. This report presents our winning solution which ctook first place in the RealADSim Workshop NVS track at ICCV 2025. To address the core challenges of street view extrapolation, we introduce a comprehensive four-stage pipeline. First, we employ a data-driven initialization strategy to generate a robust pseudo-LiDAR point cloud, avoiding local minima. Second, we inject strong geometric priors by modeling the road surface with a novel dimension-reduced SDF termed 2D-SDF. Third, we leverage a generative prior to create pseudo ground truth for extrapolated viewpoints, providing auxilary supervision. Finally, a data-driven adaptation network removes time-specific artifacts. On the RealADSim-NVS benchmark, our method achieves a final score of 0.441, ranking first among all participants.
Similar Papers
Hybrid Gaussian Splatting for Novel Urban View Synthesis
CV and Pattern Recognition
Creates new views of city streets from car videos.
AR-1-to-3: Single Image to Consistent 3D Object Generation via Next-View Prediction
CV and Pattern Recognition
Creates realistic 3D objects from a single picture.
Appreciate the View: A Task-Aware Evaluation Framework for Novel View Synthesis
CV and Pattern Recognition
Checks if computer-made pictures look real.