Scale-Invariant Robust Estimation of High-Dimensional Kronecker-Structured Matrices
By: Xiaoyu Zhang , Zhiyun Fan , Wenyang Zhang and more
High-dimensional Kronecker-structured estimation faces a conflict between non-convex scaling ambiguities and statistical robustness. The arbitrary factor scaling distorts gradient magnitudes, rendering standard fixed-threshold robust methods ineffective. We resolve this via Scaled Robust Gradient Descent (SRGD), which stabilizes optimization by de-scaling gradients before truncation. To further enforce interpretability, we introduce Scaled Hard Thresholding (SHT) for invariant variable selection. A two-step estimation procedure, built upon robust initialization and SRGD--SHT iterative updates, is proposed for canonical matrix problems, such as trace regression, matrix GLMs, and bilinear models. The convergence rates are established for heavy-tailed predictors and noise, identifying a phase transition where optimal convergence rates recover under finite noise variance and degrade optimally for heavier tails. Experiments on simulated data and two real-world applications confirm superior robustness and efficiency of the proposed procedure.
Similar Papers
Guaranteed Nonconvex Low-Rank Tensor Estimation via Scaled Gradient Descent
Machine Learning (Stat)
Cleans messy data to find hidden patterns faster.
High-dimensional low-rank matrix regression with unknown latent structures
Methodology
Finds patterns in data from many people.
Near-optimal Rank Adaptive Inference of High Dimensional Matrices
Information Theory
Finds hidden patterns in messy data.