EMMA: Scaling Mobile Manipulation via Egocentric Human Data
By: Lawrence Y. Zhu , Pranav Kuppili , Ryan Punamiya and more
Potential Business Impact:
Teaches robots to do tasks using human moves.
Scaling mobile manipulation imitation learning is bottlenecked by expensive mobile robot teleoperation. We present Egocentric Mobile MAnipulation (EMMA), an end-to-end framework training mobile manipulation policies from human mobile manipulation data with static robot data, sidestepping mobile teleoperation. To accomplish this, we co-train human full-body motion data with static robot data. In our experiments across three real-world tasks, EMMA demonstrates comparable performance to baselines trained on teleoperated mobile robot data (Mobile ALOHA), achieving higher or equivalent task performance in full task success. We find that EMMA is able to generalize to new spatial configurations and scenes, and we observe positive performance scaling as we increase the hours of human data, opening new avenues for scalable robotic learning in real-world environments. Details of this project can be found at https://ego-moma.github.io/.
Similar Papers
EgoMI: Learning Active Vision and Whole-Body Manipulation from Egocentric Human Demonstrations
Robotics
Robots learn to copy human actions better.
In-N-On: Scaling Egocentric Manipulation with in-the-wild and on-task Data
Robotics
Teaches robots to follow instructions like humans.
EMMOE: A Comprehensive Benchmark for Embodied Mobile Manipulation in Open Environments
Robotics
Robots understand and do everyday tasks from your words.