Small Symbols, Big Risks: Exploring Emoticon Semantic Confusion in Large Language Models
By: Weipeng Jiang , Xiaoyu Zhang , Juan Zhai and more
Emoticons are widely used in digital communication to convey affective intent, yet their safety implications for Large Language Models (LLMs) remain largely unexplored. In this paper, we identify emoticon semantic confusion, a vulnerability where LLMs misinterpret ASCII-based emoticons to perform unintended and even destructive actions. To systematically study this phenomenon, we develop an automated data generation pipeline and construct a dataset containing 3,757 code-oriented test cases spanning 21 meta-scenarios, four programming languages, and varying contextual complexities. Our study on six LLMs reveals that emoticon semantic confusion is pervasive, with an average confusion ratio exceeding 38%. More critically, over 90% of confused responses yield 'silent failures', which are syntactically valid outputs but deviate from user intent, potentially leading to destructive security consequences. Furthermore, we observe that this vulnerability readily transfers to popular agent frameworks, while existing prompt-based mitigations remain largely ineffective. We call on the community to recognize this emerging vulnerability and develop effective mitigation methods to uphold the safety and reliability of the LLM system.
Similar Papers
When Smiley Turns Hostile: Interpreting How Emojis Trigger LLMs' Toxicity
Computation and Language
Emojis trick computers into saying bad things.
The Hidden Language of Harm: Examining the Role of Emojis in Harmful Online Communication and Content Moderation
Computation and Language
Changes bad emojis to good ones online.
Consistency of Responses and Continuations Generated by Large Language Models on Social Media
Computation and Language
AI tones down angry online talk.