Efficient Group Lasso Regularized Rank Regression with Data-Driven Parameter Determination
By: Meixia Lin , Meijiao Shi , Yunhai Xiao and more
Potential Business Impact:
Makes computer predictions more trustworthy with bad data.
High-dimensional regression often suffers from heavy-tailed noise and outliers, which can severely undermine the reliability of least-squares based methods. To improve robustness, we adopt a non-smooth Wilcoxon score based rank objective and incorporate structured group sparsity regularization, a natural generalization of the lasso, yielding a group lasso regularized rank regression method. By extending the tuning-free parameter selection scheme originally developed for the lasso, we introduce a data-driven, simulation-based tuning rule and further establish a finite-sample error bound for the resulting estimator. On the computational side, we develop a proximal augmented Lagrangian method for solving the associated optimization problem, which eliminates the singularity issues encountered in existing methods, thereby enabling efficient semismooth Newton updates for the subproblems. Extensive numerical experiments demonstrate the robustness and effectiveness of our proposed estimator against alternatives, and showcase the scalability of the algorithm across both simulated and real-data settings.
Similar Papers
An Easily Tunable Approach to Robust and Sparse High-Dimensional Linear Regression
Statistics Theory
Finds hidden patterns even with messy data.
Regularized Reduced Rank Regression for mixed predictor and response variables
Methodology
Finds important patterns in messy, big data.
Heavy Lasso: sparse penalized regression under heavy-tailed noise via data-augmented soft-thresholding
Methodology
Makes computer models better with messy data.