Motion Focus Recognition in Fast-Moving Egocentric Video
By: Daniel Hong , James Tribble , Hao Wang and more
From Vision-Language-Action (VLA) systems to robotics, existing egocentric datasets primarily focus on action recognition tasks, while largely overlooking the inherent role of motion analysis in sports and other fast-movement scenarios. To bridge this gap, we propose a real-time motion focus recognition method that estimates the subject's locomotion intention from any egocentric video. Our approach leverages the foundation model for camera pose estimation and introduces system-level optimizations to enable efficient and scalable inference. Evaluated on a collected egocentric action dataset, our method achieves real-time performance with manageable memory consumption through a sliding batch inference strategy. This work makes motion-centric analysis practical for edge deployment and offers a complementary perspective to existing egocentric studies on sports and fast-movement activities.
Similar Papers
Ego-centric Predictive Model Conditioned on Hand Trajectories
CV and Pattern Recognition
Predicts actions and what happens next.
Ego-centric Predictive Model Conditioned on Hand Trajectories
CV and Pattern Recognition
Predicts what you'll do and what happens next.
Robust Egocentric Visual Attention Prediction Through Language-guided Scene Context-aware Learning
CV and Pattern Recognition
Helps cameras know where to look next.