Learning Human Motion with Temporally Conditional Mamba
By: Quang Nguyen , Tri Le , Baoru Huang and more
Potential Business Impact:
Makes computer-made people move like real humans.
Learning human motion based on a time-dependent input signal presents a challenging yet impactful task with various applications. The goal of this task is to generate or estimate human movement that consistently reflects the temporal patterns of conditioning inputs. Existing methods typically rely on cross-attention mechanisms to fuse the condition with motion. However, this approach primarily captures global interactions and struggles to maintain step-by-step temporal alignment. To address this limitation, we introduce Temporally Conditional Mamba, a new mamba-based model for human motion generation. Our approach integrates conditional information into the recurrent dynamics of the Mamba block, enabling better temporally aligned motion. To validate the effectiveness of our method, we evaluate it on a variety of human motion tasks. Extensive experiments demonstrate that our model significantly improves temporal alignment, motion realism, and condition consistency over state-of-the-art approaches. Our project page is available at https://zquang2202.github.io/TCM.
Similar Papers
TSkel-Mamba: Temporal Dynamic Modeling via State Space Model for Human Skeleton-based Action Recognition
CV and Pattern Recognition
Helps computers understand human movements better.
InterMamba: Efficient Human-Human Interaction Generation with Adaptive Spatio-Temporal Mamba
CV and Pattern Recognition
Makes computer-made people move together realistically.
EgoMusic-driven Human Dance Motion Estimation with Skeleton Mamba
CV and Pattern Recognition
Makes dancers move perfectly to music and video.