Non-Asymptotic Analysis of Data Augmentation for Precision Matrix Estimation
By: Lucas Morisset, Adrien Hardy, Alain Durmus
Potential Business Impact:
Helps computers learn better from more data.
This paper addresses the problem of inverse covariance (also known as precision matrix) estimation in high-dimensional settings. Specifically, we focus on two classes of estimators: linear shrinkage estimators with a target proportional to the identity matrix, and estimators derived from data augmentation (DA). Here, DA refers to the common practice of enriching a dataset with artificial samples--typically generated via a generative model or through random transformations of the original data--prior to model fitting. For both classes of estimators, we derive estimators and provide concentration bounds for their quadratic error. This allows for both method comparison and hyperparameter tuning, such as selecting the optimal proportion of artificial samples. On the technical side, our analysis relies on tools from random matrix theory. We introduce a novel deterministic equivalent for generalized resolvent matrices, accommodating dependent samples with specific structure. We support our theoretical results with numerical experiments.
Similar Papers
Scalable Bayesian inference on high-dimensional multivariate linear regression
Methodology
Finds hidden patterns in complex data.
SCOPE: Spectral Concentration by Distributionally Robust Joint Covariance-Precision Estimation
Machine Learning (Stat)
Makes computer math better for tricky data.
High-dimensional Analysis of Synthetic Data Selection
Machine Learning (Stat)
Makes fake data help computers learn better.