KeyPointDiffuser: Unsupervised 3D Keypoint Learning via Latent Diffusion Models
By: Rhys Newbury , Juyan Zhang , Tin Tran and more
Potential Business Impact:
Teaches computers to see and build 3D shapes.
Understanding and representing the structure of 3D objects in an unsupervised manner remains a core challenge in computer vision and graphics. Most existing unsupervised keypoint methods are not designed for unconditional generative settings, restricting their use in modern 3D generative pipelines; our formulation explicitly bridges this gap. We present an unsupervised framework for learning spatially structured 3D keypoints from point cloud data. These keypoints serve as a compact and interpretable representation that conditions an Elucidated Diffusion Model (EDM) to reconstruct the full shape. The learned keypoints exhibit repeatable spatial structure across object instances and support smooth interpolation in keypoint space, indicating that they capture geometric variation. Our method achieves strong performance across diverse object categories, yielding a 6 percentage-point improvement in keypoint consistency compared to prior approaches.
Similar Papers
PointDico: Contrastive 3D Representation Learning Guided by Diffusion Models
CV and Pattern Recognition
Teaches computers to understand 3D shapes better.
KineDiff3D: Kinematic-Aware Diffusion for Category-Level Articulated Object Shape Reconstruction and Generation
CV and Pattern Recognition
Builds 3D models of moving objects from one picture.
PDT: Point Distribution Transformation with Diffusion Models
CV and Pattern Recognition
Changes messy 3D shapes into neat, useful ones.