Self-Predictive Dynamics for Generalization of Vision-based Reinforcement Learning
By: Kyungsoo Kim, Jeongsoo Ha, Yusung Kim
Potential Business Impact:
Teaches robots to learn from messy pictures.
Vision-based reinforcement learning requires efficient and robust representations of image-based observations, especially when the images contain distracting (task-irrelevant) elements such as shadows, clouds, and light. It becomes more important if those distractions are not exposed during training. We design a Self-Predictive Dynamics (SPD) method to extract task-relevant features efficiently, even in unseen observations after training. SPD uses weak and strong augmentations in parallel, and learns representations by predicting inverse and forward transitions across the two-way augmented versions. In a set of MuJoCo visual control tasks and an autonomous driving task (CARLA), SPD outperforms previous studies in complex observations, and significantly improves the generalization performance for unseen observations. Our code is available at https://github.com/unigary/SPD.
Similar Papers
Self-Supervised Multisensory Pretraining for Contact-Rich Robot Reinforcement Learning
Robotics
Robots learn to grab things better with many senses.
Inverse RL Scene Dynamics Learning for Nonlinear Predictive Control in Autonomous Vehicles
Robotics
Helps self-driving cars learn better routes.
Spatial-Temporal Aware Visuomotor Diffusion Policy Learning
Robotics
Robots learn to copy actions by watching and predicting.