Breaking the Loop: Detecting and Mitigating Denial-of-Service Vulnerabilities in Large Language Models
By: Junzhe Yu , Yi Liu , Huijia Sun and more
Potential Business Impact:
Stops AI from repeating itself, making it faster.
Large Language Models (LLMs) have significantly advanced text understanding and generation, becoming integral to applications across education, software development, healthcare, entertainment, and legal services. Despite considerable progress in improving model reliability, latency remains under-explored, particularly through recurrent generation, where models repeatedly produce similar or identical outputs, causing increased latency and potential Denial-of-Service (DoS) vulnerabilities. We propose RecurrentGenerator, a black-box evolutionary algorithm that efficiently identifies recurrent generation scenarios in prominent LLMs like LLama-3 and GPT-4o. Additionally, we introduce RecurrentDetector, a lightweight real-time classifier trained on activation patterns, achieving 95.24% accuracy and an F1 score of 0.87 in detecting recurrent loops. Our methods provide practical solutions to mitigate latency-related vulnerabilities, and we publicly share our tools and data to support further research.
Similar Papers
LoopLLM: Transferable Energy-Latency Attacks in LLMs via Repetitive Generation
Cryptography and Security
Makes AI models get stuck and waste power.
ThinkTrap: Denial-of-Service Attacks against Black-box LLM Services via Infinite Thinking
Cryptography and Security
Makes AI get stuck, stopping its work.
LingoLoop Attack: Trapping MLLMs via Linguistic Context and State Entrapment into Endless Loops
Computation and Language
Makes AI models get stuck and repeat words.