DynaPose4D: High-Quality 4D Dynamic Content Generation via Pose Alignment Loss
By: Jing Yang, Yufeng Yang
Potential Business Impact:
Makes one picture move like a video.
Recent advancements in 2D and 3D generative models have expanded the capabilities of computer vision. However, generating high-quality 4D dynamic content from a single static image remains a significant challenge. Traditional methods have limitations in modeling temporal dependencies and accurately capturing dynamic geometry changes, especially when considering variations in camera perspective. To address this issue, we propose DynaPose4D, an innovative solution that integrates 4D Gaussian Splatting (4DGS) techniques with Category-Agnostic Pose Estimation (CAPE) technology. This framework uses 3D Gaussian Splatting to construct a 3D model from single images, then predicts multi-view pose keypoints based on one-shot support from a chosen view, leveraging supervisory signals to enhance motion consistency. Experimental results show that DynaPose4D achieves excellent coherence, consistency, and fluidity in dynamic motion generation. These findings not only validate the efficacy of the DynaPose4D framework but also indicate its potential applications in the domains of computer vision and animation production.
Similar Papers
PAGE-4D: Disentangled Pose and Geometry Estimation for 4D Perception
CV and Pattern Recognition
Helps cameras understand moving things in 3D.
Embracing Dynamics: Dynamics-aware 4D Gaussian Splatting SLAM
Robotics
Lets robots map moving things accurately.
MVG4D: Image Matrix-Based Multi-View and Motion Generation for 4D Content Creation from a Single Image
CV and Pattern Recognition
Makes one picture move like a video.