Shape My Moves: Text-Driven Shape-Aware Synthesis of Human Motions
By: Ting-Hsuan Liao , Yi Zhou , Yu Shen and more
Potential Business Impact:
Makes computer characters move like real people.
We explore how body shapes influence human motion synthesis, an aspect often overlooked in existing text-to-motion generation methods due to the ease of learning a homogenized, canonical body shape. However, this homogenization can distort the natural correlations between different body shapes and their motion dynamics. Our method addresses this gap by generating body-shape-aware human motions from natural language prompts. We utilize a finite scalar quantization-based variational autoencoder (FSQ-VAE) to quantize motion into discrete tokens and then leverage continuous body shape information to de-quantize these tokens back into continuous, detailed motion. Additionally, we harness the capabilities of a pretrained language model to predict both continuous shape parameters and motion tokens, facilitating the synthesis of text-aligned motions and decoding them into shape-aware motions. We evaluate our method quantitatively and qualitatively, and also conduct a comprehensive perceptual study to demonstrate its efficacy in generating shape-aware motions.
Similar Papers
Text-driven Motion Generation: Overview, Challenges and Directions
CV and Pattern Recognition
Lets computers make characters move from words.
ShapeShift: Towards Text-to-Shape Arrangement Synthesis with Content-Aware Geometric Constraints
CV and Pattern Recognition
Arranges shapes to make pictures from words.
Strong and Controllable 3D Motion Generation
CV and Pattern Recognition
Makes computer characters move faster and better.