Calib3R: A 3D Foundation Model for Multi-Camera to Robot Calibration and 3D Metric-Scaled Scene Reconstruction
By: Davide Allegro, Matteo Terreran, Stefano Ghidoni
Potential Business Impact:
Robots see in 3D without special markers.
Robots often rely on RGB images for tasks like manipulation and navigation. However, reliable interaction typically requires a 3D scene representation that is metric-scaled and aligned with the robot reference frame. This depends on accurate camera-to-robot calibration and dense 3D reconstruction, tasks usually treated separately, despite both relying on geometric correspondences from RGB data. Traditional calibration needs patterns, while RGB-based reconstruction yields geometry with an unknown scale in an arbitrary frame. Multi-camera setups add further complexity, as data must be expressed in a shared reference frame. We present Calib3R, a patternless method that jointly performs camera-to-robot calibration and metric-scaled 3D reconstruction via unified optimization. Calib3R handles single- and multi-camera setups on robot arms or mobile robots. It builds on the 3D foundation model MASt3R to extract pointmaps from RGB images, which are combined with robot poses to reconstruct a scaled 3D scene aligned with the robot. Experiments on diverse datasets show that Calib3R achieves accurate calibration with less than 10 images, outperforming target-less and marker-based methods.
Similar Papers
AMB3R: Accurate Feed-forward Metric-scale 3D Reconstruction with Backend
CV and Pattern Recognition
Creates detailed 3D models from many pictures.
Adapt3R: Adaptive 3D Scene Representation for Domain Transfer in Imitation Learning
CV and Pattern Recognition
Teaches robots to do new jobs without retraining.
Marker-Based Extrinsic Calibration Method for Accurate Multi-Camera 3D Reconstruction
CV and Pattern Recognition
Aligns 3D camera pictures perfectly for clear models.