Knockoffs Inference under Privacy Constraints
By: Zhanrui Cai, Yingying Fan, Lan Gao
Potential Business Impact:
Keeps data private while finding important information.
Model-X knockoff framework offers a model-free variable selection method that ensures finite sample false discovery rate (FDR) control. However, the complexity of generating knockoff variables, coupled with the model-free assumption, presents significant challenges for protecting data privacy in this context. In this paper, we propose a comprehensive framework for knockoff inference within the differential privacy paradigm. Our proposed method guarantees robust privacy protection while preserving the exact FDR control entailed by the original model-X knockoff procedure. We further conduct power analysis and establish sufficient conditions under which the noise added for privacy preservation does not asymptotically compromise power. Through various applications, we demonstrate that the differential privacy knockoff (DP-knockoff) method can be effectively utilized to safeguard privacy during variable selection with FDR control in both low and high dimensional settings.
Similar Papers
Gold after Randomized Sand: Model-X Split Knockoffs for Controlled Transformation Selection
Methodology
Finds important patterns in messy data.
Differentially Private Model-X Knockoffs via Johnson-Lindenstrauss Transform
Machine Learning (Stat)
Keeps private data safe while finding important patterns.
Knockoffs for low dimensions: changing the nominal level post-hoc to gain power while controlling the FDR
Methodology
Finds hidden patterns more reliably in data.