Point2Pose: A Generative Framework for 3D Human Pose Estimation with Multi-View Point Cloud Dataset
By: Hyunsoo Lee, Daeum Jeon, Hyeokjae Oh
Potential Business Impact:
Helps computers understand how people move in 3D.
We propose a novel generative approach for 3D human pose estimation. 3D human pose estimation poses several key challenges due to the complex geometry of the human body, self-occluding joints, and the requirement for large-scale real-world motion datasets. To address these challenges, we introduce Point2Pose, a framework that effectively models the distribution of human poses conditioned on sequential point cloud and pose history. Specifically, we employ a spatio-temporal point cloud encoder and a pose feature encoder to extract joint-wise features, followed by an attention-based generative regressor. Additionally, we present a large-scale indoor dataset MVPose3D, which contains multiple modalities, including IMU data of non-trivial human motions, dense multi-view point clouds, and RGB images. Experimental results show that the proposed method outperforms the baseline models, demonstrating its superior performance across various datasets.
Similar Papers
Physics Informed Human Posture Estimation Based on 3D Landmarks from Monocular RGB-Videos
CV and Pattern Recognition
Makes exercise apps understand your body better.
DreamPose3D: Hallucinative Diffusion with Prompt Learning for 3D Human Pose Estimation
CV and Pattern Recognition
Lets computers guess people's movements in 3D.
Sketch2PoseNet: Efficient and Generalized Sketch to 3D Human Pose Prediction
CV and Pattern Recognition
Draws 3D human poses from simple drawings.