PrivacyBench: A Conversational Benchmark for Evaluating Privacy in Personalized AI
By: Srija Mukhopadhyay , Sathwik Reddy , Shruthi Muthukumar and more
Personalized AI agents rely on access to a user's digital footprint, which often includes sensitive data from private emails, chats and purchase histories. Yet this access creates a fundamental societal and privacy risk: systems lacking social-context awareness can unintentionally expose user secrets, threatening digital well-being. We introduce PrivacyBench, a benchmark with socially grounded datasets containing embedded secrets and a multi-turn conversational evaluation to measure secret preservation. Testing Retrieval-Augmented Generation (RAG) assistants reveals that they leak secrets in up to 26.56% of interactions. A privacy-aware prompt lowers leakage to 5.12%, yet this measure offers only partial mitigation. The retrieval mechanism continues to access sensitive data indiscriminately, which shifts the entire burden of privacy preservation onto the generator. This creates a single point of failure, rendering current architectures unsafe for wide-scale deployment. Our findings underscore the urgent need for structural, privacy-by-design safeguards to ensure an ethical and inclusive web for everyone.
Similar Papers
PersonaBench: Evaluating AI Models on Understanding Personal Information through Accessing (Synthetic) Private User Data
Artificial Intelligence
Helps AI understand you better from your private stuff.
PrivaCI-Bench: Evaluating Privacy with Contextual Integrity and Legal Compliance
Computation and Language
Tests if AI keeps secrets safe and follows rules.
Towards Aligning Personalized Conversational Recommendation Agents with Users' Privacy Preferences
Human-Computer Interaction
AI learns your privacy rules to protect you.