Research Program: Theory of Learning in Dynamical Systems
By: Elad Hazan, Shai Shalev Shwartz, Nathan Srebro
Modern learning systems increasingly interact with data that evolve over time and depend on hidden internal state. We ask a basic question: when is such a dynamical system learnable from observations alone? This paper proposes a research program for understanding learnability in dynamical systems through the lens of next-token prediction. We argue that learnability in dynamical systems should be studied as a finite-sample question, and be based on the properties of the underlying dynamics rather than the statistical properties of the resulting sequence. To this end, we give a formulation of learnability for stochastic processes induced by dynamical systems, focusing on guarantees that hold uniformly at every time step after a finite burn-in period. This leads to a notion of dynamic learnability which captures how the structure of a system, such as stability, mixing, observability, and spectral properties, governs the number of observations required before reliable prediction becomes possible. We illustrate the framework in the case of linear dynamical systems, showing that accurate prediction can be achieved after finite observation without system identification, by leveraging improper methods based on spectral filtering. We survey the relationship between learning in dynamical systems and classical PAC, online, and universal prediction theories, and suggest directions for studying nonlinear and controlled systems.
Similar Papers
Universal Learning of Nonlinear Dynamics
Machine Learning (CS)
Learns how things change, even when wobbly.
Learning Dynamics from Infrequent Output Measurements for Uncertainty-Aware Optimal Control
Systems and Control
Controls machines even with bad, slow information.
Predicting partially observable dynamical systems via diffusion models with a multiscale inference scheme
Machine Learning (CS)
Predicts sun's future from limited views.