Dyadic Mamba: Long-term Dyadic Human Motion Synthesis
By: Julian Tanke , Takashi Shibuya , Kengo Uchida and more
Potential Business Impact:
Makes computer characters move realistically for any length.
Generating realistic dyadic human motion from text descriptions presents significant challenges, particularly for extended interactions that exceed typical training sequence lengths. While recent transformer-based approaches have shown promising results for short-term dyadic motion synthesis, they struggle with longer sequences due to inherent limitations in positional encoding schemes. In this paper, we introduce Dyadic Mamba, a novel approach that leverages State-Space Models (SSMs) to generate high-quality dyadic human motion of arbitrary length. Our method employs a simple yet effective architecture that facilitates information flow between individual motion sequences through concatenation, eliminating the need for complex cross-attention mechanisms. We demonstrate that Dyadic Mamba achieves competitive performance on standard short-term benchmarks while significantly outperforming transformer-based approaches on longer sequences. Additionally, we propose a new benchmark for evaluating long-term motion synthesis quality, providing a standardized framework for future research. Our results demonstrate that SSM-based architectures offer a promising direction for addressing the challenging task of long-term dyadic human motion synthesis from text descriptions.
Similar Papers
InterMamba: Efficient Human-Human Interaction Generation with Adaptive Spatio-Temporal Mamba
CV and Pattern Recognition
Makes computer-made people move together realistically.
TSkel-Mamba: Temporal Dynamic Modeling via State Space Model for Human Skeleton-based Action Recognition
CV and Pattern Recognition
Helps computers understand human movements better.
Trajectory Mamba: Efficient Attention-Mamba Forecasting Model Based on Selective SSM
CV and Pattern Recognition
Helps self-driving cars predict where others will go.