Alignment-Sensitive Minimax Rates for Spectral Algorithms with Learned Kernels
By: Dongming Huang , Zhifan Li , Yicheng Li and more
Potential Business Impact:
Learns better from data, improving computer predictions.
We study spectral algorithms in the setting where kernels are learned from data. We introduce the effective span dimension (ESD), an alignment-sensitive complexity measure that depends jointly on the signal, spectrum, and noise level $\sigma^2$. The ESD is well-defined for arbitrary kernels and signals without requiring eigen-decay conditions or source conditions. We prove that for sequence models whose ESD is at most $K$, the minimax excess risk scales as $\sigma^2 K$. Furthermore, we analyze over-parameterized gradient flow and prove that it can reduce the ESD. This finding establishes a connection between adaptive feature learning and provable improvements in generalization of spectral algorithms. We demonstrate the generality of the ESD framework by extending it to linear models and RKHS regression, and we support the theory with numerical experiments. This framework provides a novel perspective on generalization beyond traditional fixed-kernel theories.
Similar Papers
Alignment-Sensitive Minimax Rates for Spectral Algorithms with Learned Kernels
Machine Learning (CS)
Learns better from data, improving computer predictions.
The Minimax Lower Bound of Kernel Stein Discrepancy Estimation
Machine Learning (Stat)
Measures how well data fits, faster and better.
Spectral Concentration at the Edge of Stability: Information Geometry of Kernel Associative Memory
Machine Learning (CS)
Makes computer brains learn better and remember more.