A Fast and Light-weight Non-Iterative Visual Odometry with RGB-D Cameras
By: Zheng Yang , Kuan Xu , Shenghai Yuan and more
Potential Business Impact:
Makes robots see and move faster.
In this paper, we introduce a novel approach for efficiently estimating the 6-Degree-of-Freedom (DoF) robot pose with a decoupled, non-iterative method that capitalizes on overlapping planar elements. Conventional RGB-D visual odometry(RGBD-VO) often relies on iterative optimization solvers to estimate pose and involves a process of feature extraction and matching. This results in significant computational burden and time delays. To address this, our innovative method for RGBD-VO separates the estimation of rotation and translation. Initially, we exploit the overlaid planar characteristics within the scene to calculate the rotation matrix. Following this, we utilize a kernel cross-correlator (KCC) to ascertain the translation. By sidestepping the resource-intensive iterative optimization and feature extraction and alignment procedures, our methodology offers improved computational efficacy, achieving a performance of 71Hz on a lower-end i5 CPU. When the RGBD-VO does not rely on feature points, our technique exhibits enhanced performance in low-texture degenerative environments compared to state-of-the-art methods.
Similar Papers
360DVO: Deep Visual Odometry for Monocular 360-Degree Camera
CV and Pattern Recognition
Helps robots see where they are going better.
6-DoF Object Tracking with Event-based Optical Flow and Frames
CV and Pattern Recognition
Tracks fast-moving objects with special cameras.
SMF-VO: Direct Ego-Motion Estimation via Sparse Motion Fields
Robotics
Lets robots see where they are going faster.