ConceptPose: Training-Free Zero-Shot Object Pose Estimation using Concept Vectors
By: Liming Kuang , Yordanka Velikova , Mahdi Saleh and more
Object pose estimation is a fundamental task in computer vision and robotics, yet most methods require extensive, dataset-specific training. Concurrently, large-scale vision language models show remarkable zero-shot capabilities. In this work, we bridge these two worlds by introducing ConceptPose, a framework for object pose estimation that is both training-free and model-free. ConceptPose leverages a vision-language-model (VLM) to create open-vocabulary 3D concept maps, where each point is tagged with a concept vector derived from saliency maps. By establishing robust 3D-3D correspondences across concept maps, our approach allows precise estimation of 6DoF relative pose. Without any object or dataset-specific training, our approach achieves state-of-the-art results on common zero shot relative pose estimation benchmarks, significantly outperforming existing methods by over 62% in ADD(-S) score, including those that utilize extensive dataset-specific training.
Similar Papers
Beyond 'Templates': Category-Agnostic Object Pose, Size, and Shape Estimation from a Single View
CV and Pattern Recognition
Helps robots understand and grab any object.
Articulate3D: Zero-Shot Text-Driven 3D Object Posing
CV and Pattern Recognition
Moves 3D objects with just words.
ViTa-Zero: Zero-shot Visuotactile Object 6D Pose Estimation
Robotics
Robots can better grab and move objects.