DexCtrl: Towards Sim-to-Real Dexterity with Adaptive Controller Learning
By: Shuqi Zhao , Ke Yang , Yuxin Chen and more
Potential Business Impact:
Robots learn to do tasks better in real life.
Dexterous manipulation has seen remarkable progress in recent years, with policies capable of executing many complex and contact-rich tasks in simulation. However, transferring these policies from simulation to real world remains a significant challenge. One important issue is the mismatch in low-level controller dynamics, where identical trajectories can lead to vastly different contact forces and behaviors when control parameters vary. Existing approaches often rely on manual tuning or controller randomization, which can be labor-intensive, task-specific, and introduce significant training difficulty. In this work, we propose a framework that jointly learns actions and controller parameters based on the historical information of both trajectory and controller. This adaptive controller adjustment mechanism allows the policy to automatically tune control parameters during execution, thereby mitigating the sim-to-real gap without extensive manual tuning or excessive randomization. Moreover, by explicitly providing controller parameters as part of the observation, our approach facilitates better reasoning over force interactions and improves robustness in real-world scenarios. Experimental results demonstrate that our method achieves improved transfer performance across a variety of dexterous tasks involving variable force conditions.
Similar Papers
DexterityGen: Foundation Controller for Unprecedented Dexterity
Robotics
Robots learn to use tools like humans.
DexTrack: Towards Generalizable Neural Tracking Control for Dexterous Manipulation from Human References
Robotics
Robot hands learn to copy human movements.
Dexterous Manipulation Transfer via Progressive Kinematic-Dynamic Alignment
Robotics
Robots copy human hand moves from videos.