PALUM: Part-based Attention Learning for Unified Motion Retargeting
By: Siqi Liu , Maoyu Wang , Bo Dai and more
Retargeting motion between characters with different skeleton structures is a fundamental challenge in computer animation. When source and target characters have vastly different bone arrangements, maintaining the original motion's semantics and quality becomes increasingly difficult. We present PALUM, a novel approach that learns common motion representations across diverse skeleton topologies by partitioning joints into semantic body parts and applying attention mechanisms to capture spatio-temporal relationships. Our method transfers motion to target skeletons by leveraging these skeleton-agnostic representations alongside target-specific structural information. To ensure robust learning and preserve motion fidelity, we introduce a cycle consistency mechanism that maintains semantic coherence throughout the retargeting process. Extensive experiments demonstrate superior performance in handling diverse skeletal structures while maintaining motion realism and semantic fidelity, even when generalizing to previously unseen skeleton-motion combinations. We will make our implementation publicly available to support future research.
Similar Papers
Skeletons Speak Louder than Text: A Motion-Aware Pretraining Paradigm for Video-Based Person Re-Identification
CV and Pattern Recognition
Helps computers recognize people in videos by their movement.
World-Coordinate Human Motion Retargeting via SAM 3D Body
Robotics
Makes robots copy human movements from videos.
SUGAR: Learning Skeleton Representation with Visual-Motion Knowledge for Action Recognition
CV and Pattern Recognition
Teaches computers to understand human movements.