Minimal Deterministic Echo State Networks Outperform Random Reservoirs in Learning Chaotic Dynamics
By: Francesco Martinuzzi
Potential Business Impact:
Makes computers better at predicting messy things.
Machine learning (ML) is widely used to model chaotic systems. Among ML approaches, echo state networks (ESNs) have received considerable attention due to their simple construction and fast training. However, ESN performance is highly sensitive to hyperparameter choices and to its random initialization. In this work, we demonstrate that ESNs constructed using deterministic rules and simple topologies (MESNs) outperform standard ESNs in the task of chaotic attractor reconstruction. We use a dataset of more than 90 chaotic systems to benchmark 10 different minimal deterministic reservoir initializations. We find that MESNs obtain up to a 41% reduction in error compared to standard ESNs. Furthermore, we show that the MESNs are more robust, exhibiting less inter-run variation, and have the ability to reuse hyperparameters across different systems. Our results illustrate how structured simplicity in ESN design can outperform stochastic complexity in learning chaotic dynamics.
Similar Papers
Echo State Networks as State-Space Models: A Systems Perspective
Machine Learning (CS)
Makes smart computers learn faster and better.
HypER: Hyperbolic Echo State Networks for Capturing Stretch-and-Fold Dynamics in Chaotic Flows
Machine Learning (CS)
Predicts chaotic events much further into the future.
Contraction, Criticality, and Capacity: A Dynamical-Systems Perspective on Echo-State Networks
Neural and Evolutionary Computing
Makes computers remember past information better.