Nonlocal Neural Tangent Kernels via Parameter-Space Interactions
By: Sriram Nagaraj, Vishakh Hari
Potential Business Impact:
Helps computers learn from messy, imperfect data.
The Neural Tangent Kernel (NTK) framework has provided deep insights into the training dynamics of neural networks under gradient flow. However, it relies on the assumption that the network is differentiable with respect to its parameters, an assumption that breaks down when considering non-smooth target functions or parameterized models exhibiting non-differentiable behavior. In this work, we propose a Nonlocal Neural Tangent Kernel (NNTK) that replaces the local gradient with a nonlocal interaction-based approximation in parameter space. Nonlocal gradients are known to exist for a wider class of functions than the standard gradient. This allows NTK theory to be extended to nonsmooth functions, stochastic estimators, and broader families of models. We explore both fixed-kernel and attention-based formulations of this nonlocal operator. We illustrate the new formulation with numerical studies.
Similar Papers
Neural Tangent Kernels for Complex Genetic Risk Prediction: Bridging Deep Learning and Kernel Methods in Genomics
Applications
Finds hidden disease risks in your genes.
Mathematical Foundations of Neural Tangents and Infinite-Width Networks
Machine Learning (CS)
Makes AI learn better and faster.
Finite-Width Neural Tangent Kernels from Feynman Diagrams
Machine Learning (CS)
Helps computers learn better by understanding tiny changes.