Prediction with Expert Advice under Local Differential Privacy
By: Ben Jacobsen, Kassem Fawaz
Potential Business Impact:
Protects patient data while predicting disease spread.
We study the classic problem of prediction with expert advice under the constraint of local differential privacy (LDP). In this context, we first show that a classical algorithm naturally satisfies LDP and then design two new algorithms that improve it: RW-AdaBatch and RW-Meta. For RW-AdaBatch, we exploit the limited-switching behavior induced by LDP to provide a novel form of privacy amplification that grows stronger on easier data, analogous to the shuffle model in offline learning. Drawing on the theory of random walks, we prove that this improvement carries essentially no utility cost. For RW-Meta, we develop a general method for privately selecting between experts that are themselves non-trivial learning algorithms, and we show that in the context of LDP this carries no extra privacy cost. In contrast, prior work has only considered data-independent experts. We also derive formal regret bounds that scale inversely with the degree of independence between experts. Our analysis is supplemented by evaluation on real-world data reported by hospitals during the COVID-19 pandemic; RW-Meta outperforms both the classical baseline and a state-of-the-art \textit{central} DP algorithm by 1.5-3$\times$ on the task of predicting which hospital will report the highest density of COVID patients each week.
Similar Papers
High-Probability Bounds For Heterogeneous Local Differential Privacy
Machine Learning (Stat)
Protects your private info while still getting useful data.
Privacy-Utility-Bias Trade-offs for Privacy-Preserving Recommender Systems
Machine Learning (CS)
Protects user data while still suggesting good movies.
Optimal Fairness under Local Differential Privacy
Machine Learning (CS)
Makes private data fair for computers.