AMB3R: Accurate Feed-forward Metric-scale 3D Reconstruction with Backend
By: Hengyi Wang, Lourdes Agapito
Potential Business Impact:
Creates detailed 3D models from many pictures.
We present AMB3R, a multi-view feed-forward model for dense 3D reconstruction on a metric-scale that addresses diverse 3D vision tasks. The key idea is to leverage a sparse, yet compact, volumetric scene representation as our backend, enabling geometric reasoning with spatial compactness. Although trained solely for multi-view reconstruction, we demonstrate that AMB3R can be seamlessly extended to uncalibrated visual odometry (online) or large-scale structure from motion without the need for task-specific fine-tuning or test-time optimization. Compared to prior pointmap-based models, our approach achieves state-of-the-art performance in camera pose, depth, and metric-scale estimation, 3D reconstruction, and even surpasses optimization-based SLAM and SfM methods with dense reconstruction priors on common benchmarks.
Similar Papers
Any4D: Unified Feed-Forward Metric 4D Reconstruction
CV and Pattern Recognition
Makes videos show moving 3D objects accurately.
SAB3R: Semantic-Augmented Backbone in 3D Reconstruction
CV and Pattern Recognition
Lets computers build 3D maps from videos.
Calib3R: A 3D Foundation Model for Multi-Camera to Robot Calibration and 3D Metric-Scaled Scene Reconstruction
Robotics
Robots see in 3D without special markers.