Uni-Inter: Unifying 3D Human Motion Synthesis Across Diverse Interaction Contexts
By: Sheng Liu , Yuanzhi Liang , Jiepeng Wang and more
Potential Business Impact:
Makes computer characters move realistically together.
We present Uni-Inter, a unified framework for human motion generation that supports a wide range of interaction scenarios: including human-human, human-object, and human-scene-within a single, task-agnostic architecture. In contrast to existing methods that rely on task-specific designs and exhibit limited generalization, Uni-Inter introduces the Unified Interactive Volume (UIV), a volumetric representation that encodes heterogeneous interactive entities into a shared spatial field. This enables consistent relational reasoning and compound interaction modeling. Motion generation is formulated as joint-wise probabilistic prediction over the UIV, allowing the model to capture fine-grained spatial dependencies and produce coherent, context-aware behaviors. Experiments across three representative interaction tasks demonstrate that Uni-Inter achieves competitive performance and generalizes well to novel combinations of entities. These results suggest that unified modeling of compound interactions offers a promising direction for scalable motion synthesis in complex environments.
Similar Papers
UniMo: Unifying 2D Video and 3D Human Motion with an Autoregressive Framework
CV and Pattern Recognition
Creates matching 3D moves from videos.
InterSyn: Interleaved Learning for Dynamic Motion Synthesis in the Wild
CV and Pattern Recognition
Makes computer characters move together realistically.
UniEgoMotion: A Unified Model for Egocentric Motion Reconstruction, Forecasting, and Generation
CV and Pattern Recognition
Lets computers guess how people will move.