On Optimal Hyperparameters for Differentially Private Deep Transfer Learning
By: Aki Rehn , Linzh Zhao , Mikko A. Heikkilä and more
Potential Business Impact:
Protects private data while training smart computer programs.
Differentially private (DP) transfer learning, i.e., fine-tuning a pretrained model on private data, is the current state-of-the-art approach for training large models under privacy constraints. We focus on two key hyperparameters in this setting: the clipping bound $C$ and batch size $B$. We show a clear mismatch between the current theoretical understanding of how to choose an optimal $C$ (stronger privacy requires smaller $C$) and empirical outcomes (larger $C$ performs better under strong privacy), caused by changes in the gradient distributions. Assuming a limited compute budget (fixed epochs), we demonstrate that the existing heuristics for tuning $B$ do not work, while cumulative DP noise better explains whether smaller or larger batches perform better. We also highlight how the common practice of using a single $(C,B)$ setting across tasks can lead to suboptimal performance. We find that performance drops especially when moving between loose and tight privacy and between plentiful and limited compute, which we explain by analyzing clipping as a form of gradient re-weighting and examining cumulative DP noise.
Similar Papers
Mitigating Disparate Impact of Differentially Private Learning through Bounded Adaptive Clipping
Machine Learning (CS)
Protects privacy without hurting fairness for all.
Towards hyperparameter-free optimization with differential privacy
Machine Learning (CS)
Trains AI privately without needing to test many settings.
Parameter-Efficient Fine-Tuning with Differential Privacy for Robust Instruction Adaptation in Large Language Models
Computation and Language
Keeps AI learning private and fast.