Learning global control of underactuated systems with Model-Based Reinforcement Learning
By: Niccolò Turcato , Marco Calì , Alberto Dalla Libera and more
Potential Business Impact:
Teaches robots to learn new tasks faster.
This short paper describes our proposed solution for the third edition of the "AI Olympics with RealAIGym" competition, held at ICRA 2025. We employed Monte-Carlo Probabilistic Inference for Learning Control (MC-PILCO), an MBRL algorithm recognized for its exceptional data efficiency across various low-dimensional robotic tasks, including cart-pole, ball \& plate, and Furuta pendulum systems. MC-PILCO optimizes a system dynamics model using interaction data, enabling policy refinement through simulation rather than direct system data optimization. This approach has proven highly effective in physical systems, offering greater data efficiency than Model-Free (MF) alternatives. Notably, MC-PILCO has previously won the first two editions of this competition, demonstrating its robustness in both simulated and real-world environments. Besides briefly reviewing the algorithm, we discuss the most critical aspects of the MC-PILCO implementation in the tasks at hand: learning a global policy for the pendubot and acrobot systems.
Similar Papers
Accelerating Model-Based Reinforcement Learning using Non-Linear Trajectory Optimization
Machine Learning (CS)
Teaches robots new skills much faster.
Reinforcement Learning for Robust Athletic Intelligence: Lessons from the 2nd 'AI Olympics with RealAIGym' Competition
Robotics
Robots learn to balance wobbly things better.
Model Identification Adaptive Control with $ρ$-POMDP Planning
Robotics
Teaches robots to learn and control things better.