OpenEgo: A Large-Scale Multimodal Egocentric Dataset for Dexterous Manipulation
By: Ahad Jawaid, Yu Xiang
Potential Business Impact:
Teaches robots to copy human hand movements.
Egocentric human videos provide scalable demonstrations for imitation learning, but existing corpora often lack either fine-grained, temporally localized action descriptions or dexterous hand annotations. We introduce OpenEgo, a multimodal egocentric manipulation dataset with standardized hand-pose annotations and intention-aligned action primitives. OpenEgo totals 1107 hours across six public datasets, covering 290 manipulation tasks in 600+ environments. We unify hand-pose layouts and provide descriptive, timestamped action primitives. To validate its utility, we train language-conditioned imitation-learning policies to predict dexterous hand trajectories. OpenEgo is designed to lower the barrier to learning dexterous manipulation from egocentric video and to support reproducible research in vision-language-action learning. All resources and instructions will be released at www.openegocentric.com.
Similar Papers
EgoDex: Learning Dexterous Manipulation from Large-Scale Egocentric Video
CV and Pattern Recognition
Teaches robots to do tasks by watching humans.
IndEgo: A Dataset of Industrial Scenarios and Collaborative Work for Egocentric Assistants
CV and Pattern Recognition
Helps robots learn to do factory jobs.
MultiEgo: A Multi-View Egocentric Video Dataset for 4D Scene Reconstruction
CV and Pattern Recognition
Records real-life events from multiple viewpoints.