Privacy Risk Predictions Based on Fundamental Understanding of Personal Data and an Evolving Threat Landscape
By: Haoran Niu, K. Suzanne Barber
Potential Business Impact:
Shows how one data leak can cause others.
It is difficult for individuals and organizations to protect personal information without a fundamental understanding of relative privacy risks. By analyzing over 5,000 empirical identity theft and fraud cases, this research identifies which types of personal data are exposed, how frequently exposures occur, and what the consequences of those exposures are. We construct an Identity Ecosystem graph--a foundational, graph-based model in which nodes represent personally identifiable information (PII) attributes and edges represent empirical disclosure relationships between them (e.g., the probability that one PII attribute is exposed due to the exposure of another). Leveraging this graph structure, we develop a privacy risk prediction framework that uses graph theory and graph neural networks to estimate the likelihood of further disclosures when certain PII attributes are compromised. The results show that our approach effectively answers the core question: Can the disclosure of a given identity attribute possibly lead to the disclosure of another attribute?
Similar Papers
How Worrying Are Privacy Attacks Against Machine Learning?
Cryptography and Security
Protects personal information used to train AI.
Understanding Privacy Risks in Code Models Through Training Dynamics: A Causal Approach
Software Engineering
Protects private code details from AI learning.
Understanding Privacy Risks in Code Models Through Training Dynamics: A Causal Approach
Software Engineering
Protects private code secrets from AI.