Score: 0

Learning from Less: SINDy Surrogates in RL

Published: April 25, 2025 | arXiv ID: 2504.18113v1

By: Aniket Dixit , Muhammad Ibrahim Khan , Faizan Ahmed and more

Potential Business Impact:

Teaches robots faster with less practice.

Business Areas:
Simulation Software

This paper introduces an approach for developing surrogate environments in reinforcement learning (RL) using the Sparse Identification of Nonlinear Dynamics (SINDy) algorithm. We demonstrate the effectiveness of our approach through extensive experiments in OpenAI Gym environments, particularly Mountain Car and Lunar Lander. Our results show that SINDy-based surrogate models can accurately capture the underlying dynamics of these environments while reducing computational costs by 20-35%. With only 75 interactions for Mountain Car and 1000 for Lunar Lander, we achieve state-wise correlations exceeding 0.997, with mean squared errors as low as 3.11e-06 for Mountain Car velocity and 1.42e-06 for LunarLander position. RL agents trained in these surrogate environments require fewer total steps (65,075 vs. 100,000 for Mountain Car and 801,000 vs. 1,000,000 for Lunar Lander) while achieving comparable performance to those trained in the original environments, exhibiting similar convergence patterns and final performance metrics. This work contributes to the field of model-based RL by providing an efficient method for generating accurate, interpretable surrogate environments.

Country of Origin
🇬🇧 United Kingdom

Page Count
6 pages

Category
Computer Science:
Machine Learning (CS)