Fundamental limits of distributed covariance matrix estimation via a conditional strong data processing inequality
By: Mohammad Reza Rahmani, Mohammad Hossein Yassaee, Mohammad Reza Aref
Potential Business Impact:
Lets computers share data safely for better guessing.
Estimating high-dimensional covariance matrices is a key task across many fields. This paper explores the theoretical limits of distributed covariance estimation in a feature-split setting, where communication between agents is constrained. Specifically, we study a scenario in which multiple agents each observe different components of i.i.d. samples drawn from a sub-Gaussian random vector. A central server seeks to estimate the complete covariance matrix using a limited number of bits communicated by each agent. We obtain a nearly tight minimax lower bound for covariance matrix estimation under operator norm and Frobenius norm. Our main technical tool is a novel generalization of the strong data processing inequality (SDPI), termed the Conditional Strong Data Processing Inequality (C-SDPI) coefficient, introduced in this work. The C-SDPI coefficient shares key properties such as tensorization with the conventional SDPI. Crucially, it quantifies the average contraction in a state-dependent channel and can be significantly lower than the worst-case SDPI coefficient over the state input. Utilizing the doubling trick of Geng-Nair and an operator Jensen inequality, we compute this coefficient for Gaussian mixture channels. We then employ it to establish minimax lower bounds on estimation error, capturing the trade-offs among sample size, communication cost, and data dimensionality. Building on this, we present a nearly optimal estimation protocol whose sample and communication requirements match the lower bounds up to logarithmic factors. Unlike much of the existing literature, our framework does not assume infinite samples or Gaussian distributions, making it broadly applicable. Finally, we extend our analysis to interactive protocols, showing interaction can significantly reduce communication requirements compared to non-interactive schemes.
Similar Papers
SCOPE: Spectral Concentration by Distributionally Robust Joint Covariance-Precision Estimation
Machine Learning (Stat)
Makes computer math better for tricky data.
A DPI-PAC-Bayesian Framework for Generalization Bounds
Information Theory
Makes computer learning more accurate and reliable.
Non-Linear Strong Data-Processing for Quantum Hockey-Stick Divergences
Quantum Physics
Makes secret messages safer in quantum computers.