Decentralized Quantile Regression for Feature-Distributed Massive Datasets with Privacy Guarantees
By: Peiwen Xiao , Xiaohui Liu , Guangming Pan and more
Potential Business Impact:
Protects private data while learning from many computers.
In this paper, we introduce a novel decentralized surrogate gradient-based algorithm for quantile regression in a feature-distributed setting, where global features are dispersed across multiple machines within a decentralized network. The proposed algorithm, \texttt{DSG-cqr}, utilizes a convolution-type smoothing approach to address the non-smooth nature of the quantile loss function. \texttt{DSG-cqr} is fully decentralized, conjugate-free, easy to implement, and achieves linear convergence up to statistical precision. To ensure privacy, we adopt the Gaussian mechanism to provide $(\epsilon,\delta)$-differential privacy. To overcome the exact residual calculation problem, we estimate residuals using auxiliary variables and develop a confidence interval construction method based on Wald statistics. Theoretical properties are established, and the practical utility of the methods is also demonstrated through extensive simulations and a real-world data application.
Similar Papers
High-Dimensional Differentially Private Quantile Regression: Distributed Estimation and Statistical Inference
Machine Learning (Stat)
Keeps your private data safe when analyzing it.
Federated Learning of Quantile Inference under Local Differential Privacy
Methodology
Helps computers learn from private data safely.
Sparse-Smooth Spatially Varying Coefficient Quantile Regression
Methodology
Finds how things change from place to place.