PoseSyn: Synthesizing Diverse 3D Pose Data from In-the-Wild 2D Data
By: ChangHee Yang , Hyeonseop Song , Seokhun Choi and more
Potential Business Impact:
Creates realistic 3D poses for better computer vision.
Despite considerable efforts to enhance the generalization of 3D pose estimators without costly 3D annotations, existing data augmentation methods struggle in real world scenarios with diverse human appearances and complex poses. We propose PoseSyn, a novel data synthesis framework that transforms abundant in the wild 2D pose dataset into diverse 3D pose image pairs. PoseSyn comprises two key components: Error Extraction Module (EEM), which identifies challenging poses from the 2D pose datasets, and Motion Synthesis Module (MSM), which synthesizes motion sequences around the challenging poses. Then, by generating realistic 3D training data via a human animation model aligned with challenging poses and appearances PoseSyn boosts the accuracy of various 3D pose estimators by up to 14% across real world benchmarks including various backgrounds and occlusions, challenging poses, and multi view scenarios. Extensive experiments further confirm that PoseSyn is a scalable and effective approach for improving generalization without relying on expensive 3D annotations, regardless of the pose estimator's model size or design.
Similar Papers
SEPose: A Synthetic Event-based Human Pose Estimation Dataset for Pedestrian Monitoring
CV and Pattern Recognition
Helps cameras see people in bad conditions.
Point2Pose: A Generative Framework for 3D Human Pose Estimation with Multi-View Point Cloud Dataset
CV and Pattern Recognition
Helps computers understand how people move in 3D.
Synthetic Human Action Video Data Generation with Pose Transfer
CV and Pattern Recognition
Makes fake videos of people move realistically.