Can Context Bridge the Reality Gap? Sim-to-Real Transfer of Context-Aware Policies
By: Marco Iannotta , Yuxuan Yang , Johannes A. Stork and more
Potential Business Impact:
Robots learn better from computer games.
Sim-to-real transfer remains a major challenge in reinforcement learning (RL) for robotics, as policies trained in simulation often fail to generalize to the real world due to discrepancies in environment dynamics. Domain Randomization (DR) mitigates this issue by exposing the policy to a wide range of randomized dynamics during training, yet leading to a reduction in performance. While standard approaches typically train policies agnostic to these variations, we investigate whether sim-to-real transfer can be improved by conditioning the policy on an estimate of the dynamics parameters -- referred to as context. To this end, we integrate a context estimation module into a DR-based RL framework and systematically compare SOTA supervision strategies. We evaluate the resulting context-aware policies in both a canonical control benchmark and a real-world pushing task using a Franka Emika Panda robot. Results show that context-aware policies outperform the context-agnostic baseline across all settings, although the best supervision strategy depends on the task.
Similar Papers
Sim-to-Real Transfer in Deep Reinforcement Learning for Bipedal Locomotion
Robotics
Robots learn to walk outside the computer.
Bridging the Reality Gap in Digital Twins with Context-Aware, Physics-Guided Deep Learning
Machine Learning (CS)
Makes digital copies of things stay accurate.
The Reality Gap in Robotics: Challenges, Solutions, and Best Practices
Robotics
Helps robots learn in games, then work in real life.