The Siren Song of LLMs: How Users Perceive and Respond to Dark Patterns in Large Language Models
By: Yike Shi , Qing Xiao , Qing Hu and more
Potential Business Impact:
AI tricks people with fake helpfulness in chats.
Large language models can influence users through conversation, creating new forms of dark patterns that differ from traditional UX dark patterns. We define LLM dark patterns as manipulative or deceptive behaviors enacted in dialogue. Drawing on prior work and AI incident reports, we outline a diverse set of categories with real-world examples. Using them, we conducted a scenario-based study where participants (N=34) compared manipulative and neutral LLM responses. Our results reveal that recognition of LLM dark patterns often hinged on conversational cues such as exaggerated agreement, biased framing, or privacy intrusions, but these behaviors were also sometimes normalized as ordinary assistance. Users' perceptions of these dark patterns shaped how they respond to them. Responsibilities for these behaviors were also attributed in different ways, with participants assigning it to companies and developers, the model itself, or to users. We conclude with implications for design, advocacy, and governance to safeguard user autonomy.
Similar Papers
The Siren Song of LLMs: How Users Perceive and Respond to Dark Patterns in Large Language Models
Human-Computer Interaction
AI tricks people with fake helpfulness in chats.
DarkBench: Benchmarking Dark Patterns in Large Language Models
Computation and Language
Finds sneaky tricks in AI that trick people.
Investigating the Impact of Dark Patterns on LLM-Based Web Agents
Cryptography and Security
Protects online helpers from tricky website tricks.