Skyfall-GS: Synthesizing Immersive 3D Urban Scenes from Satellite Imagery
By: Jie-Ying Lee , Yi-Ruei Liu , Shr-Ruei Tsai and more
Potential Business Impact:
Builds 3D cities from satellite pictures.
Synthesizing large-scale, explorable, and geometrically accurate 3D urban scenes is a challenging yet valuable task in providing immersive and embodied applications. The challenges lie in the lack of large-scale and high-quality real-world 3D scans for training generalizable generative models. In this paper, we take an alternative route to create large-scale 3D scenes by synergizing the readily available satellite imagery that supplies realistic coarse geometry and the open-domain diffusion model for creating high-quality close-up appearances. We propose \textbf{Skyfall-GS}, the first city-block scale 3D scene creation framework without costly 3D annotations, also featuring real-time, immersive 3D exploration. We tailor a curriculum-driven iterative refinement strategy to progressively enhance geometric completeness and photorealistic textures. Extensive experiments demonstrate that Skyfall-GS provides improved cross-view consistent geometry and more realistic textures compared to state-of-the-art approaches. Project page: https://skyfall-gs.jayinnn.dev/
Similar Papers
From Orbit to Ground: Generative City Photogrammetry from Extreme Off-Nadir Satellite Images
CV and Pattern Recognition
Creates 3D city views from satellite pictures.
Satellite to GroundScape -- Large-scale Consistent Ground View Generation from Satellite Views
CV and Pattern Recognition
Turns bird's-eye views into connected street scenes.
MetroGS: Efficient and Stable Reconstruction of Geometrically Accurate High-Fidelity Large-Scale Scenes
CV and Pattern Recognition
Builds detailed 3D city maps from photos.