What Can One Expect When Solving PDEs Using Shallow Neural Networks?
By: Roy Y. He , Ying Liang , Hongkai Zhao and more
Potential Business Impact:
Makes computers solve hard math problems better.
We use elliptic partial differential equations (PDEs) as examples to show various properties and behaviors when shallow neural networks (SNNs) are used to represent the solutions. In particular, we study the numerical ill-conditioning, frequency bias, and the balance between the differential operator and the shallow network representation for different formulations of the PDEs and with various activation functions. Our study shows that the performance of Physics-Informed Neural Networks (PINNs) or Deep Ritz Method (DRM) using linear SNNs with power ReLU activation is dominated by their inherent ill-conditioning and spectral bias against high frequencies. Although this can be alleviated by using non-homogeneous activation functions with proper scaling, achieving such adaptivity for nonlinear SNNs remains costly due to ill-conditioning.
Similar Papers
What Can One Expect When Solving PDEs Using Shallow Neural Networks?
Numerical Analysis
Neural networks learn math problems better.
Separated-Variable Spectral Neural Networks: A Physics-Informed Learning Approach for High-Frequency PDEs
Machine Learning (CS)
Solves wiggly physics equations 1000 times better
Deep vs. Shallow: Benchmarking Physics-Informed Neural Architectures on the Biharmonic Equation
Computational Engineering, Finance, and Science
Makes computer simulations of science faster.