Generalizable Domain Adaptation for Sim-and-Real Policy Co-Training
By: Shuo Cheng , Liqian Ma , Zhenyang Chen and more
Potential Business Impact:
Teaches robots to do tasks with less real practice.
Behavior cloning has shown promise for robot manipulation, but real-world demonstrations are costly to acquire at scale. While simulated data offers a scalable alternative, particularly with advances in automated demonstration generation, transferring policies to the real world is hampered by various simulation and real domain gaps. In this work, we propose a unified sim-and-real co-training framework for learning generalizable manipulation policies that primarily leverages simulation and only requires a few real-world demonstrations. Central to our approach is learning a domain-invariant, task-relevant feature space. Our key insight is that aligning the joint distributions of observations and their corresponding actions across domains provides a richer signal than aligning observations (marginals) alone. We achieve this by embedding an Optimal Transport (OT)-inspired loss within the co-training framework, and extend this to an Unbalanced OT framework to handle the imbalance between abundant simulation data and limited real-world examples. We validate our method on challenging manipulation tasks, showing it can leverage abundant simulation data to achieve up to a 30% improvement in the real-world success rate and even generalize to scenarios seen only in simulation.
Similar Papers
Safe Continual Domain Adaptation after Sim2Real Transfer of Reinforcement Learning Policies in Robotics
Robotics
Lets robots learn and change safely in the real world.
Train Once, Deploy Anywhere: Realize Data-Efficient Dynamic Object Manipulation
Robotics
Robots learn to grab many things with few examples.
X-Sim: Cross-Embodiment Learning via Real-to-Sim-to-Real
Robotics
Teaches robots to do tasks from human videos.