Score: 0

3D sans 3D Scans: Scalable Pre-training from Video-Generated Point Clouds

Published: December 28, 2025 | arXiv ID: 2512.23042v1

By: Ryousuke Yamada , Kohsuke Ide , Yoshihiro Fukuhara and more

Potential Business Impact:

Teaches computers 3D shapes from regular videos.

Business Areas:
Image Recognition Data and Analytics, Software

Despite recent progress in 3D self-supervised learning, collecting large-scale 3D scene scans remains expensive and labor-intensive. In this work, we investigate whether 3D representations can be learned from unlabeled videos recorded without any real 3D sensors. We present Laplacian-Aware Multi-level 3D Clustering with Sinkhorn-Knopp (LAM3C), a self-supervised framework that learns from video-generated point clouds from unlabeled videos. We first introduce RoomTours, a video-generated point cloud dataset constructed by collecting room-walkthrough videos from the web (e.g., real-estate tours) and generating 49,219 scenes using an off-the-shelf feed-forward reconstruction model. We also propose a noise-regularized loss that stabilizes representation learning by enforcing local geometric smoothness and ensuring feature stability under noisy point clouds. Remarkably, without using any real 3D scans, LAM3C achieves higher performance than the previous self-supervised methods on indoor semantic and instance segmentation. These results suggest that unlabeled videos represent an abundant source of data for 3D self-supervised learning.

Page Count
15 pages

Category
Computer Science:
CV and Pattern Recognition