Ego-Exo 3D Hand Tracking in the Wild with a Mobile Multi-Camera Rig
By: Patrick Rim , Kun He , Kevin Harris and more
Potential Business Impact:
Tracks hands in 3D, even when moving freely.
Accurate 3D tracking of hands and their interactions with the world in unconstrained settings remains a significant challenge for egocentric computer vision. With few exceptions, existing datasets are predominantly captured in controlled lab setups, limiting environmental diversity and model generalization. To address this, we introduce a novel marker-less multi-camera system designed to capture precise 3D hands and objects, which allows for nearly unconstrained mobility in genuinely in-the-wild conditions. We combine a lightweight, back-mounted capture rig with eight exocentric cameras, and a user-worn Meta Quest 3 headset, which contributes two egocentric views. We design an ego-exo tracking pipeline to generate accurate 3D hand pose ground truth from this system, and rigorously evaluate its quality. By collecting an annotated dataset featuring synchronized multi-view images and precise 3D hand poses, we demonstrate the capability of our approach to significantly reduce the trade-off between environmental realism and 3D annotation accuracy.
Similar Papers
Uni-Hand: Universal Hand Motion Forecasting in Egocentric Views
CV and Pattern Recognition
Finds exact moments hands touch objects.
The Invisible EgoHand: 3D Hand Forecasting through EgoBody Pose Estimation
CV and Pattern Recognition
Predicts where hands will move, even when hidden.
OpenEgo: A Large-Scale Multimodal Egocentric Dataset for Dexterous Manipulation
CV and Pattern Recognition
Teaches robots to copy human hand movements.