Score: 0

Kernel ridge regression under power-law data: spectrum and generalization

Published: October 6, 2025 | arXiv ID: 2510.04780v1

By: Arie Wortsman, Bruno Loureiro

Potential Business Impact:

Helps computers learn better from messy data.

Business Areas:
A/B Testing Data and Analytics

In this work, we investigate high-dimensional kernel ridge regression (KRR) on i.i.d. Gaussian data with anisotropic power-law covariance. This setting differs fundamentally from the classical source & capacity conditions for KRR, where power-law assumptions are typically imposed on the kernel eigen-spectrum itself. Our contributions are twofold. First, we derive an explicit characterization of the kernel spectrum for polynomial inner-product kernels, giving a precise description of how the kernel eigen-spectrum inherits the data decay. Second, we provide an asymptotic analysis of the excess risk in the high-dimensional regime for a particular kernel with this spectral behavior, showing that the sample complexity is governed by the effective dimension of the data rather than the ambient dimension. These results establish a fundamental advantage of learning with power-law anisotropic data over isotropic data. To our knowledge, this is the first rigorous treatment of non-linear KRR under power-law data.

Page Count
38 pages

Category
Statistics:
Machine Learning (Stat)