Heavy Lasso: sparse penalized regression under heavy-tailed noise via data-augmented soft-thresholding
By: The Tien Mai
Potential Business Impact:
Makes computer models better with messy data.
High-dimensional linear regression is a fundamental tool in modern statistics, particularly when the number of predictors exceeds the sample size. The classical Lasso, which relies on the squared loss, performs well under Gaussian noise assumptions but often deteriorates in the presence of heavy-tailed errors or outliers commonly encountered in real data applications such as genomics, finance, and signal processing. To address these challenges, we propose a novel robust regression method, termed Heavy Lasso, which incorporates a loss function inspired by the Student's t-distribution within a Lasso penalization framework. This loss retains the desirable quadratic behavior for small residuals while adaptively downweighting large deviations, thus enhancing robustness to heavy-tailed noise and outliers. Heavy Lasso enjoys computationally efficient by leveraging a data augmentation scheme and a soft-thresholding algorithm, which integrate seamlessly with classical Lasso solvers. Theoretically, we establish non-asymptotic bounds under both $\ell_1$ and $\ell_2 $ norms, by employing the framework of localized convexity, showing that the Heavy Lasso estimator achieves rates comparable to those of the Huber loss. Extensive numerical studies demonstrate Heavy Lasso's superior performance over classical Lasso and other robust variants, highlighting its effectiveness in challenging noisy settings. Our method is implemented in the R package heavylasso available on Github.
Similar Papers
Exponential Lasso: robust sparse penalization under heavy-tailed noise and outliers with exponential-type loss
Machine Learning (Stat)
Makes computer models ignore bad data points.
An Easily Tunable Approach to Robust and Sparse High-Dimensional Linear Regression
Statistics Theory
Finds hidden patterns even with messy data.
Understanding Robust Machine Learning for Nonparametric Regression with Heavy-Tailed Noise
Machine Learning (CS)
Makes computers learn from messy, unreliable data.