Unleashing the Temporal Potential of Stereo Event Cameras for Continuous-Time 3D Object Detection
By: Jae-Young Kang, Hoonhee Cho, Kuk-Jin Yoon
Potential Business Impact:
Lets self-driving cars see moving objects better.
3D object detection is essential for autonomous systems, enabling precise localization and dimension estimation. While LiDAR and RGB cameras are widely used, their fixed frame rates create perception gaps in high-speed scenarios. Event cameras, with their asynchronous nature and high temporal resolution, offer a solution by capturing motion continuously. The recent approach, which integrates event cameras with conventional sensors for continuous-time detection, struggles in fast-motion scenarios due to its dependency on synchronized sensors. We propose a novel stereo 3D object detection framework that relies solely on event cameras, eliminating the need for conventional 3D sensors. To compensate for the lack of semantic and geometric information in event data, we introduce a dual filter mechanism that extracts both. Additionally, we enhance regression by aligning bounding boxes with object-centric information. Experiments show that our method outperforms prior approaches in dynamic environments, demonstrating the potential of event cameras for robust, continuous-time 3D perception. The code is available at https://github.com/mickeykang16/Ev-Stereo3D.
Similar Papers
A Survey of 3D Reconstruction with Event Cameras
CV and Pattern Recognition
Helps robots see in fast, dark, or bright places.
Event Camera Guided Visual Media Restoration & 3D Reconstruction: A Survey
CV and Pattern Recognition
Improves blurry videos and 3D pictures.
6-DoF Object Tracking with Event-based Optical Flow and Frames
CV and Pattern Recognition
Tracks fast-moving objects with special cameras.