Lower Bounds for Public-Private Learning under Distribution Shift
By: Amrith Setlur, Pratiksha Thaker, Jonathan Ullman
Potential Business Impact:
Combines data to learn better, but only if it's similar.
The most effective differentially private machine learning algorithms in practice rely on an additional source of purportedly public data. This paradigm is most interesting when the two sources combine to be more than the sum of their parts. However, there are settings such as mean estimation where we have strong lower bounds, showing that when the two data sources have the same distribution, there is no complementary value to combining the two data sources. In this work we extend the known lower bounds for public-private learning to setting where the two data sources exhibit significant distribution shift. Our results apply to both Gaussian mean estimation where the two distributions have different means, and to Gaussian linear regression where the two distributions exhibit parameter shift. We find that when the shift is small (relative to the desired accuracy), either public or private data must be sufficiently abundant to estimate the private parameter. Conversely, when the shift is large, public data provides no benefit.
Similar Papers
ModShift: Model Privacy via Designed Shifts
Machine Learning (CS)
Keeps computer learning private from spies.
Shift is Good: Mismatched Data Mixing Improves Test Performance
Machine Learning (CS)
Makes AI learn better with mixed-up practice.
Bounding User Contributions for User-Level Differentially Private Mean Estimation
Information Theory
Keeps private data safe when sharing averages.