MonoPlace3D: Learning 3D-Aware Object Placement for 3D Monocular Detection
By: Rishubh Parihar , Srinjay Sarkar , Sarthak Vora and more
Potential Business Impact:
Makes self-driving cars see better in 3D.
Current monocular 3D detectors are held back by the limited diversity and scale of real-world datasets. While data augmentation certainly helps, it's particularly difficult to generate realistic scene-aware augmented data for outdoor settings. Most current approaches to synthetic data generation focus on realistic object appearance through improved rendering techniques. However, we show that where and how objects are positioned is just as crucial for training effective 3D monocular detectors. The key obstacle lies in automatically determining realistic object placement parameters - including position, dimensions, and directional alignment when introducing synthetic objects into actual scenes. To address this, we introduce MonoPlace3D, a novel system that considers the 3D scene content to create realistic augmentations. Specifically, given a background scene, MonoPlace3D learns a distribution over plausible 3D bounding boxes. Subsequently, we render realistic objects and place them according to the locations sampled from the learned distribution. Our comprehensive evaluation on two standard datasets KITTI and NuScenes, demonstrates that MonoPlace3D significantly improves the accuracy of multiple existing monocular 3D detectors while being highly data efficient.
Similar Papers
MonoLite3D: Lightweight 3D Object Properties Estimation
CV and Pattern Recognition
Helps self-driving cars see in 3D from one camera.
PlaceIt3D: Language-Guided Object Placement in Real 3D Scenes
CV and Pattern Recognition
Places 3D objects in scenes using words.
MonoCT: Overcoming Monocular 3D Detection Domain Shift with Consistent Teacher Models
CV and Pattern Recognition
Helps cars see in 3D without extra cameras.