OnlineSplatter: Pose-Free Online 3D Reconstruction for Free-Moving Objects
By: Mark He Huang , Lin Geng Foo , Christian Theobalt and more
Potential Business Impact:
Makes 3D models of moving things from videos.
Free-moving object reconstruction from monocular video remains challenging, particularly without reliable pose or depth cues and under arbitrary object motion. We introduce OnlineSplatter, a novel online feed-forward framework generating high-quality, object-centric 3D Gaussians directly from RGB frames without requiring camera pose, depth priors, or bundle optimization. Our approach anchors reconstruction using the first frame and progressively refines the object representation through a dense Gaussian primitive field, maintaining constant computational cost regardless of video sequence length. Our core contribution is a dual-key memory module combining latent appearance-geometry keys with explicit directional keys, robustly fusing current frame features with temporally aggregated object states. This design enables effective handling of free-moving objects via spatial-guided memory readout and an efficient sparsification mechanism, ensuring comprehensive yet compact object coverage. Evaluations on real-world datasets demonstrate that OnlineSplatter significantly outperforms state-of-the-art pose-free reconstruction baselines, consistently improving with more observations while maintaining constant memory and runtime.
Similar Papers
Online 3D Gaussian Splatting Modeling with Novel View Selection
CV and Pattern Recognition
Creates more complete 3D models from fewer pictures.
Online 3D Gaussian Splatting Modeling with Novel View Selection
CV and Pattern Recognition
Makes 3D pictures more complete from videos.
Pseudo Depth Meets Gaussian: A Feed-forward RGB SLAM Baseline
CV and Pattern Recognition
Makes 3D models from videos much faster.