Seeing My Future: Predicting Situated Interaction Behavior in Virtual Reality
By: Yuan Xu , Zimu Zhang , Xiaoxuan Ma and more
Potential Business Impact:
Predicts what you'll do next in virtual worlds.
Virtual and augmented reality systems increasingly demand intelligent adaptation to user behaviors for enhanced interaction experiences. Achieving this requires accurately understanding human intentions and predicting future situated behaviors - such as gaze direction and object interactions - which is vital for creating responsive VR/AR environments and applications like personalized assistants. However, accurate behavioral prediction demands modeling the underlying cognitive processes that drive human-environment interactions. In this work, we introduce a hierarchical, intention-aware framework that models human intentions and predicts detailed situated behaviors by leveraging cognitive mechanisms. Given historical human dynamics and the observation of scene contexts, our framework first identifies potential interaction targets and forecasts fine-grained future behaviors. We propose a dynamic Graph Convolutional Network (GCN) to effectively capture human-environment relationships. Extensive experiments on challenging real-world benchmarks and live VR environment demonstrate the effectiveness of our approach, achieving superior performance across all metrics and enabling practical applications for proactive VR systems that anticipate user behaviors and adapt virtual environments accordingly.
Similar Papers
Predicting User Grasp Intentions in Virtual Reality
Human-Computer Interaction
Makes VR hands move like real hands.
CSGaze: Context-aware Social Gaze Prediction
CV and Pattern Recognition
Helps computers understand where people are looking.
LookOut: Real-World Humanoid Egocentric Navigation
CV and Pattern Recognition
Helps robots and computers understand where you're looking.