Next-Generation Reservoir Computing for Dynamical Inference
By: Rok Cestnik, Erik A. Martens
Potential Business Impact:
Models complex systems from messy data.
We present a simple and scalable implementation of next-generation reservoir computing for modeling dynamical systems from time series data. Our approach uses a pseudorandom nonlinear projection of time-delay embedded input, allowing an arbitrary dimension of the feature space, thus providing a flexible alternative to the polynomial-based projections used in previous next-generation reservoir computing variants. We apply the method to benchmark tasks -- including attractor reconstruction and bifurcation diagram estimation -- using only partial and noisy observations. We also include an exploratory example of estimating asymptotic oscillation phases. The models remain stable over long rollouts and generalize beyond training data. This framework enables the precise control of system state and is well suited for surrogate modeling and digital twin applications.
Similar Papers
Learning Beyond Experience: Generalizing to Unseen State Space with Reservoir Computing
Machine Learning (CS)
Lets computers predict unseen future events.
Next-generation reservoir computing validated by classification task
Machine Learning (CS)
Teaches computers to sort and predict information.
Cross-Scale Reservoir Computing for large spatio-temporal forecasting and modeling
Machine Learning (CS)
Predicts ocean temperatures much better, longer.