ProDyG: Progressive Dynamic Scene Reconstruction via Gaussian Splatting from Monocular Videos
By: Shi Chen , Erik Sandström , Sandro Lombardi and more
Potential Business Impact:
Builds 3D worlds from videos in real-time.
Achieving truly practical dynamic 3D reconstruction requires online operation, global pose and map consistency, detailed appearance modeling, and the flexibility to handle both RGB and RGB-D inputs. However, existing SLAM methods typically merely remove the dynamic parts or require RGB-D input, while offline methods are not scalable to long video sequences, and current transformer-based feedforward methods lack global consistency and appearance details. To this end, we achieve online dynamic scene reconstruction by disentangling the static and dynamic parts within a SLAM system. The poses are tracked robustly with a novel motion masking strategy, and dynamic parts are reconstructed leveraging a progressive adaptation of a Motion Scaffolds graph. Our method yields novel view renderings competitive to offline methods and achieves on-par tracking with state-of-the-art dynamic SLAM methods.
Similar Papers
Dy3DGS-SLAM: Monocular 3D Gaussian Splatting SLAM for Dynamic Environments
CV and Pattern Recognition
Maps moving things using only one camera.
4D Gaussian Splatting SLAM
CV and Pattern Recognition
Makes videos look real, even with moving things.
Pseudo Depth Meets Gaussian: A Feed-forward RGB SLAM Baseline
CV and Pattern Recognition
Makes 3D models from videos much faster.