Towards Better Attribute Inference Vulnerability Measures
By: Paul Francis, David Wagner
Potential Business Impact:
Protects private info while keeping data useful.
The purpose of anonymizing structured data is to protect the privacy of individuals in the data while retaining the statistical properties of the data. An important class of attack on anonymized data is attribute inference, where an attacker infers the value of an unknown attribute of a target individual given knowledge of one or more known attributes. A major limitation of recent attribute inference measures is that they do not take recall into account, only precision. It is often the case that attacks target only a fraction of individuals, for instance data outliers. Incorporating recall, however, substantially complicates the measure, because one must determine how to combine recall and precision in a composite measure for both the attack and baseline. This paper presents the design and implementation of an attribute inference measure that incorporates both precision and recall. Our design also improves on how the baseline attribute inference is computed. In experiments using a generic best row match attack on moderately-anonymized microdata, we show that in over 25\% of the attacks, our approach correctly labeled the attack to be at risk while the prior approach incorrectly labeled the attack to be safe.
Similar Papers
Disparate Privacy Vulnerability: Targeted Attribute Inference Attacks and Defenses
Machine Learning (CS)
Protects private data from sneaky computer guesses.
Multi-Objective Optimization-Based Anonymization of Structured Data for Machine Learning Application
Machine Learning (CS)
Protects private data while keeping it useful.
Inference Attacks for X-Vector Speaker Anonymization
Cryptography and Security
Keeps your voice private from sneaky listeners.