Prior-Enhanced Gaussian Splatting for Dynamic Scene Reconstruction from Casual Video
By: Meng-Li Shih , Ying-Huan Chen , Yu-Lun Liu and more
Potential Business Impact:
Makes videos look real, like you're there.
We introduce a fully automatic pipeline for dynamic scene reconstruction from casually captured monocular RGB videos. Rather than designing a new scene representation, we enhance the priors that drive Dynamic Gaussian Splatting. Video segmentation combined with epipolar-error maps yields object-level masks that closely follow thin structures; these masks (i) guide an object-depth loss that sharpens the consistent video depth, and (ii) support skeleton-based sampling plus mask-guided re-identification to produce reliable, comprehensive 2-D tracks. Two additional objectives embed the refined priors in the reconstruction stage: a virtual-view depth loss removes floaters, and a scaffold-projection loss ties motion nodes to the tracks, preserving fine geometry and coherent motion. The resulting system surpasses previous monocular dynamic scene reconstruction methods and delivers visibly superior renderings
Similar Papers
On-the-fly Large-scale 3D Reconstruction from Multi-Camera Rigs
CV and Pattern Recognition
Builds 3D worlds from many cameras fast.
MOSAIC-GS: Monocular Scene Reconstruction via Advanced Initialization for Complex Dynamic Environments
CV and Pattern Recognition
Makes videos look real by tracking moving objects.
ProDyG: Progressive Dynamic Scene Reconstruction via Gaussian Splatting from Monocular Videos
CV and Pattern Recognition
Builds 3D worlds from videos in real-time.