Novel Object 6D Pose Estimation with a Single Reference View
By: Jian Liu , Wei Sun , Kai Zeng and more
Potential Business Impact:
Lets robots find objects with just one picture.
Existing novel object 6D pose estimation methods typically rely on CAD models or dense reference views, which are both difficult to acquire. Using only a single reference view is more scalable, but challenging due to large pose discrepancies and limited geometric and spatial information. To address these issues, we propose a Single-Reference-based novel object 6D (SinRef-6D) pose estimation method. Our key idea is to iteratively establish point-wise alignment in a common coordinate system based on state space models (SSMs). Specifically, iterative object-space point-wise alignment can effectively handle large pose discrepancies, while our proposed RGB and Points SSMs can capture long-range dependencies and spatial information from a single view, offering linear complexity and superior spatial modeling capability. Once pre-trained on synthetic data, SinRef-6D can estimate the 6D pose of a novel object using only a single reference view, without requiring retraining or a CAD model. Extensive experiments on six popular datasets and real-world robotic scenes demonstrate that we achieve on-par performance with CAD-based and dense reference view-based methods, despite operating in the more challenging single reference setting. Code will be released at https://github.com/CNJianLiu/SinRef-6D.
Similar Papers
Any6D: Model-free 6D Pose Estimation of Novel Objects
CV and Pattern Recognition
Lets robots find and grab any object perfectly.
RefPose: Leveraging Reference Geometric Correspondences for Accurate 6D Pose Estimation of Unseen Objects
CV and Pattern Recognition
Helps robots find and grab new things.
One2Any: One-Reference 6D Pose Estimation for Any Object
CV and Pattern Recognition
Lets robots see objects from any angle.