Beyond PII: How Users Attempt to Estimate and Mitigate Implicit LLM Inference
By: Synthia Wang , Sai Teja Peddinti , Nina Taft and more
Potential Business Impact:
AI can guess your secrets from your words.
Large Language Models (LLMs) such as ChatGPT can infer personal attributes from seemingly innocuous text, raising privacy risks beyond memorized data leakage. While prior work has demonstrated these risks, little is known about how users estimate and respond. We conducted a survey with 240 U.S. participants who judged text snippets for inference risks, reported concern levels, and attempted rewrites to block inference. We compared their rewrites with those generated by ChatGPT and Rescriber, a state-of-the-art sanitization tool. Results show that participants struggled to anticipate inference, performing a little better than chance. User rewrites were effective in just 28\% of cases - better than Rescriber but worse than ChatGPT. We examined our participants' rewriting strategies, and observed that while paraphrasing was the most common strategy it is also the least effective; instead abstraction and adding ambiguity were more successful. Our work highlights the importance of inference-aware design in LLM interactions.
Similar Papers
Out-of-Context Abduction: LLMs Make Inferences About Procedural Data Leveraging Declarative Facts in Earlier Training Data
Computation and Language
AI guesses chatbot names from reply styles
Ask ChatGPT: Caveats and Mitigations for Individual Users of AI Chatbots
Computers and Society
Warns about AI chatbots hurting your brain.
Conversational User-AI Intervention: A Study on Prompt Rewriting for Improved LLM Response Generation
Computation and Language
Helps computers understand what you want better.