Learning from Less: SINDy Surrogates in RL
By: Aniket Dixit , Muhammad Ibrahim Khan , Faizan Ahmed and more
Potential Business Impact:
Teaches robots faster with less practice.
This paper introduces an approach for developing surrogate environments in reinforcement learning (RL) using the Sparse Identification of Nonlinear Dynamics (SINDy) algorithm. We demonstrate the effectiveness of our approach through extensive experiments in OpenAI Gym environments, particularly Mountain Car and Lunar Lander. Our results show that SINDy-based surrogate models can accurately capture the underlying dynamics of these environments while reducing computational costs by 20-35%. With only 75 interactions for Mountain Car and 1000 for Lunar Lander, we achieve state-wise correlations exceeding 0.997, with mean squared errors as low as 3.11e-06 for Mountain Car velocity and 1.42e-06 for LunarLander position. RL agents trained in these surrogate environments require fewer total steps (65,075 vs. 100,000 for Mountain Car and 801,000 vs. 1,000,000 for Lunar Lander) while achieving comparable performance to those trained in the original environments, exhibiting similar convergence patterns and final performance metrics. This work contributes to the field of model-based RL by providing an efficient method for generating accurate, interpretable surrogate environments.
Similar Papers
Dyna-Style Reinforcement Learning Modeling and Control of Non-linear Dynamics
Systems and Control
Teaches robots to learn faster and better.
Sparse identification of nonlinear dynamics with high accuracy and reliability under noisy conditions for applications to industrial systems
Systems and Control
Predicts complex engine behavior accurately, even with noise.
Sequential Regression Learning with Randomized Algorithms
Machine Learning (Stat)
Learns patterns in changing information.