Robustness in Large Language Models: A Survey of Mitigation Strategies and Evaluation Metrics
By: Pankaj Kumar, Subhankar Mishra
Potential Business Impact:
Makes smart computer programs more reliable and trustworthy.
Large Language Models (LLMs) have emerged as a promising cornerstone for the development of natural language processing (NLP) and artificial intelligence (AI). However, ensuring the robustness of LLMs remains a critical challenge. To address these challenges and advance the field, this survey provides a comprehensive overview of current studies in this area. First, we systematically examine the nature of robustness in LLMs, including its conceptual foundations, the importance of consistent performance across diverse inputs, and the implications of failure modes in real-world applications. Next, we analyze the sources of non-robustness, categorizing intrinsic model limitations, data-driven vulnerabilities, and external adversarial factors that compromise reliability. Following this, we review state-of-the-art mitigation strategies, and then we discuss widely adopted benchmarks, emerging metrics, and persistent gaps in assessing real-world reliability. Finally, we synthesize findings from existing surveys and interdisciplinary studies to highlight trends, unresolved issues, and pathways for future research.
Similar Papers
Evaluating and Improving Robustness in Large Language Models: A Survey and Future Directions
Computation and Language
Makes AI smarter and more reliable.
Attack and defense techniques in large language models: A survey and new perspectives
Cryptography and Security
Protects smart computer programs from being tricked.
Survey of Adversarial Robustness in Multimodal Large Language Models
CV and Pattern Recognition
Makes AI understand pictures and words safely.