3D sans 3D Scans: Scalable Pre-training from Video-Generated Point Clouds
By: Ryousuke Yamada , Kohsuke Ide , Yoshihiro Fukuhara and more
Potential Business Impact:
Teaches computers 3D shapes from regular videos.
Despite recent progress in 3D self-supervised learning, collecting large-scale 3D scene scans remains expensive and labor-intensive. In this work, we investigate whether 3D representations can be learned from unlabeled videos recorded without any real 3D sensors. We present Laplacian-Aware Multi-level 3D Clustering with Sinkhorn-Knopp (LAM3C), a self-supervised framework that learns from video-generated point clouds from unlabeled videos. We first introduce RoomTours, a video-generated point cloud dataset constructed by collecting room-walkthrough videos from the web (e.g., real-estate tours) and generating 49,219 scenes using an off-the-shelf feed-forward reconstruction model. We also propose a noise-regularized loss that stabilizes representation learning by enforcing local geometric smoothness and ensuring feature stability under noisy point clouds. Remarkably, without using any real 3D scans, LAM3C achieves higher performance than the previous self-supervised methods on indoor semantic and instance segmentation. These results suggest that unlabeled videos represent an abundant source of data for 3D self-supervised learning.
Similar Papers
QueryOcc: Query-based Self-Supervision for 3D Semantic Occupancy
CV and Pattern Recognition
Teaches cars to see and understand 3D worlds.
CASL: Curvature-Augmented Self-supervised Learning for 3D Anomaly Detection
CV and Pattern Recognition
Finds hidden flaws in 3D objects using shape.
Integrating SAM Supervision for 3D Weakly Supervised Point Cloud Segmentation
CV and Pattern Recognition
Helps computers understand 3D shapes with less 3D data.