Universal Adversarial Suffixes for Language Models Using Reinforcement Learning with Calibrated Reward
By: Sampriti Soor, Suklav Ghosh, Arijit Sur
Potential Business Impact:
Makes AI models easily tricked by short text.
Language models are vulnerable to short adversarial suffixes that can reliably alter predictions. Previous works usually find such suffixes with gradient search or rule-based methods, but these are brittle and often tied to a single task or model. In this paper, a reinforcement learning framework is used where the suffix is treated as a policy and trained with Proximal Policy Optimization against a frozen model as a reward oracle. Rewards are shaped using calibrated cross-entropy, removing label bias and aggregating across surface forms to improve transferability. The proposed method is evaluated on five diverse NLP benchmark datasets, covering sentiment, natural language inference, paraphrase, and commonsense reasoning, using three distinct language models: Qwen2-1.5B Instruct, TinyLlama-1.1B Chat, and Phi-1.5. Results show that RL-trained suffixes consistently degrade accuracy and transfer more effectively across tasks and models than previous adversarial triggers of similar genres.
Similar Papers
Universal Adversarial Suffixes Using Calibrated Gumbel-Softmax Relaxation
Computation and Language
Makes AI models easily fooled by bad words.
Universal and Transferable Adversarial Attack on Large Language Models Using Exponentiated Gradient Descent
Machine Learning (CS)
Stops smart computers from being tricked.
RIVAL: Reinforcement Learning with Iterative and Adversarial Optimization for Machine Translation
Computation and Language
Makes movie subtitles sound natural in any language.