OPENTOUCH: Bringing Full-Hand Touch to Real-World Interaction
By: Yuxin Ray Song , Jinzhou Li , Rao Fu and more
The human hand is our primary interface to the physical world, yet egocentric perception rarely knows when, where, or how forcefully it makes contact. Robust wearable tactile sensors are scarce, and no existing in-the-wild datasets align first-person video with full-hand touch. To bridge the gap between visual perception and physical interaction, we present OpenTouch, the first in-the-wild egocentric full-hand tactile dataset, containing 5.1 hours of synchronized video-touch-pose data and 2,900 curated clips with detailed text annotations. Using OpenTouch, we introduce retrieval and classification benchmarks that probe how touch grounds perception and action. We show that tactile signals provide a compact yet powerful cue for grasp understanding, strengthen cross-modal alignment, and can be reliably retrieved from in-the-wild video queries. By releasing this annotated vision-touch-pose dataset and benchmark, we aim to advance multimodal egocentric perception, embodied learning, and contact-rich robotic manipulation.
Similar Papers
OSMO: Open-Source Tactile Glove for Human-to-Robot Skill Transfer
Robotics
Robots learn to do tasks by feeling with gloves.
Uni-Hand: Universal Hand Motion Forecasting in Egocentric Views
CV and Pattern Recognition
Finds exact moments hands touch objects.
A Humanoid Visual-Tactile-Action Dataset for Contact-Rich Manipulation
Robotics
Robots learn to touch and grab soft things.