Score: 0

The Interplay of Statistics and Noisy Optimization: Learning Linear Predictors with Random Data Weights

Published: December 11, 2025 | arXiv ID: 2512.10188v1

By: Gabriel Clara, Yazan Mash'al

Potential Business Impact:

Makes computer learning faster and more accurate.

Business Areas:
Predictive Analytics Artificial Intelligence, Data and Analytics, Software

We analyze gradient descent with randomly weighted data points in a linear regression model, under a generic weighting distribution. This includes various forms of stochastic gradient descent, importance sampling, but also extends to weighting distributions with arbitrary continuous values, thereby providing a unified framework to analyze the impact of various kinds of noise on the training trajectory. We characterize the implicit regularization induced through the random weighting, connect it with weighted linear regression, and derive non-asymptotic bounds for convergence in first and second moments. Leveraging geometric moment contraction, we also investigate the stationary distribution induced by the added noise. Based on these results, we discuss how specific choices of weighting distribution influence both the underlying optimization problem and statistical properties of the resulting estimator, as well as some examples for which weightings that lead to fast convergence cause bad statistical performance.

Page Count
34 pages

Category
Statistics:
Machine Learning (Stat)