Learning Generalizable Visuomotor Policy through Dynamics-Alignment
By: Dohyeok Lee , Jung Min Lee , Munkyung Kim and more
Potential Business Impact:
Teaches robots to learn from mistakes better.
Behavior cloning methods for robot learning suffer from poor generalization due to limited data support beyond expert demonstrations. Recent approaches leveraging video prediction models have shown promising results by learning rich spatiotemporal representations from large-scale datasets. However, these models learn action-agnostic dynamics that cannot distinguish between different control inputs, limiting their utility for precise manipulation tasks and requiring large pretraining datasets. We propose a Dynamics-Aligned Flow Matching Policy (DAP) that integrates dynamics prediction into policy learning. Our method introduces a novel architecture where policy and dynamics models provide mutual corrective feedback during action generation, enabling self-correction and improved generalization. Empirical validation demonstrates generalization performance superior to baseline methods on real-world robotic manipulation tasks, showing particular robustness in OOD scenarios including visual distractions and lighting variations.
Similar Papers
3D Flow Diffusion Policy: Visuomotor Policy Learning via Generating Flow in 3D Space
Robotics
Robots learn to grab and move things better.
A Study on Enhancing the Generalization Ability of Visuomotor Policies via Data Augmentation
Robotics
Teaches robots to do tasks in new places.
Learning on the Fly: Rapid Policy Adaptation via Differentiable Simulation
Robotics
Robots learn to fix mistakes instantly in real world.