Satellite to GroundScape -- Large-scale Consistent Ground View Generation from Satellite Views
By: Ningli Xu, Rongjun Qin
Potential Business Impact:
Turns bird's-eye views into connected street scenes.
Generating consistent ground-view images from satellite imagery is challenging, primarily due to the large discrepancies in viewing angles and resolution between satellite and ground-level domains. Previous efforts mainly concentrated on single-view generation, often resulting in inconsistencies across neighboring ground views. In this work, we propose a novel cross-view synthesis approach designed to overcome these challenges by ensuring consistency across ground-view images generated from satellite views. Our method, based on a fixed latent diffusion model, introduces two conditioning modules: satellite-guided denoising, which extracts high-level scene layout to guide the denoising process, and satellite-temporal denoising, which captures camera motion to maintain consistency across multiple generated views. We further contribute a large-scale satellite-ground dataset containing over 100,000 perspective pairs to facilitate extensive ground scene or video generation. Experimental results demonstrate that our approach outperforms existing methods on perceptual and temporal metrics, achieving high photorealism and consistency in multi-view outputs.
Similar Papers
SatDreamer360: Geometry Consistent Street-View Video Generation from Satellite Imagery
CV and Pattern Recognition
Makes satellite pictures look like real street videos.
From Satellite to Street: A Hybrid Framework Integrating Stable Diffusion and PanoGAN for Consistent Cross-View Synthesis
CV and Pattern Recognition
Makes street pictures from satellite maps.
From Orbit to Ground: Generative City Photogrammetry from Extreme Off-Nadir Satellite Images
CV and Pattern Recognition
Creates 3D city views from satellite pictures.