SAM 3D for 3D Object Reconstruction from Remote Sensing Images
By: Junsheng Yao, Lichao Mou, Qingyu Li
Potential Business Impact:
Builds 3D city models from single photos.
Monocular 3D building reconstruction from remote sensing imagery is essential for scalable urban modeling, yet existing methods often require task-specific architectures and intensive supervision. This paper presents the first systematic evaluation of SAM 3D, a general-purpose image-to-3D foundation model, for monocular remote sensing building reconstruction. We benchmark SAM 3D against TRELLIS on samples from the NYC Urban Dataset, employing Frechet Inception Distance (FID) and CLIP-based Maximum Mean Discrepancy (CMMD) as evaluation metrics. Experimental results demonstrate that SAM 3D produces more coherent roof geometry and sharper boundaries compared to TRELLIS. We further extend SAM 3D to urban scene reconstruction through a segment-reconstruct-compose pipeline, demonstrating its potential for urban scene modeling. We also analyze practical limitations and discuss future research directions. These findings provide practical guidance for deploying foundation models in urban 3D reconstruction and motivate future integration of scene-level structural priors.
Similar Papers
SAM 3D: 3Dfy Anything in Images
CV and Pattern Recognition
Turns flat pictures into 3D objects.
SegEarth-OV3: Exploring SAM 3 for Open-Vocabulary Semantic Segmentation in Remote Sensing Images
CV and Pattern Recognition
Lets computers find any object in satellite pictures.
Ref-SAM3D: Bridging SAM3D with Text for Reference 3D Reconstruction
CV and Pattern Recognition
Makes 3D models from text and one picture.