Fourier Neural Operators for Non-Markovian Processes:Approximation Theorems and Experiments
By: Wonjae Lee, Taeyoung Kim, Hyungbin Park
Potential Business Impact:
Learns how random things change much faster.
This paper introduces an operator-based neural network, the mirror-padded Fourier neural operator (MFNO), designed to learn the dynamics of stochastic systems. MFNO extends the standard Fourier neural operator (FNO) by incorporating mirror padding, enabling it to handle non-periodic inputs. We rigorously prove that MFNOs can approximate solutions of path-dependent stochastic differential equations and Lipschitz transformations of fractional Brownian motions to an arbitrary degree of accuracy. Our theoretical analysis builds on Wong--Zakai type theorems and various approximation techniques. Empirically, the MFNO exhibits strong resolution generalization--a property rarely seen in standard architectures such as LSTMs, TCNs, and DeepONet. Furthermore, our model achieves performance that is comparable or superior to these baselines while offering significantly faster sample path generation than classical numerical schemes.
Similar Papers
Fourier Neural Operators Explained: A Practical Perspective
Machine Learning (CS)
Teaches computers to solve hard math problems faster.
Learning Function-to-Function Mappings: A Fourier Neural Operator for Next-Generation MIMO Systems
Information Theory
Makes faster internet by understanding radio waves better.
Universal Fourier Neural Operators for Micromechanics
Computational Engineering, Finance, and Science
Solves tiny material puzzles faster than before.