Leveraging Automatic CAD Annotations for Supervised Learning in 3D Scene Understanding
By: Yuchen Rao , Stefan Ainetter , Sinisa Stekovic and more
Potential Business Impact:
Teaches computers to understand 3D objects better.
High-level 3D scene understanding is essential in many applications. However, the challenges of generating accurate 3D annotations make development of deep learning models difficult. We turn to recent advancements in automatic retrieval of synthetic CAD models, and show that data generated by such methods can be used as high-quality ground truth for training supervised deep learning models. More exactly, we employ a pipeline akin to the one previously used to automatically annotate objects in ScanNet scenes with their 9D poses and CAD models. This time, we apply it to the recent ScanNet++ v1 dataset, which previously lacked such annotations. Our findings demonstrate that it is not only possible to train deep learning models on these automatically-obtained annotations but that the resulting models outperform those trained on manually annotated data. We validate this on two distinct tasks: point cloud completion and single-view CAD model retrieval and alignment. Our results underscore the potential of automatic 3D annotations to enhance model performance while significantly reducing annotation costs. To support future research in 3D scene understanding, we will release our annotations, which we call SCANnotate++, along with our trained models.
Similar Papers
Towards Auto-Annotation from Annotation Guidelines: A Benchmark through 3D LiDAR Detection
CV and Pattern Recognition
Teaches computers to label data automatically.
DenseAnnotate: Enabling Scalable Dense Caption Collection for Images and 3D Scenes via Spoken Descriptions
CV and Pattern Recognition
Lets computers understand images by listening.
3D Can Be Explored In 2D: Pseudo-Label Generation for LiDAR Point Clouds Using Sensor-Intensity-Based 2D Semantic Segmentation
CV and Pattern Recognition
Teaches self-driving cars to see without 3D maps.