Safety in Large Reasoning Models: A Survey
By: Cheng Wang , Yue Liu , Baolong Bi and more
Potential Business Impact:
Makes smart computer programs safer to use.
Large Reasoning Models (LRMs) have exhibited extraordinary prowess in tasks like mathematics and coding, leveraging their advanced reasoning capabilities. Nevertheless, as these capabilities progress, significant concerns regarding their vulnerabilities and safety have arisen, which can pose challenges to their deployment and application in real-world settings. This paper presents a comprehensive survey of LRMs, meticulously exploring and summarizing the newly emerged safety risks, attacks, and defense strategies. By organizing these elements into a detailed taxonomy, this work aims to offer a clear and structured understanding of the current safety landscape of LRMs, facilitating future research and development to enhance the security and reliability of these powerful models.
Similar Papers
The Hidden Risks of Large Reasoning Models: A Safety Assessment of R1
Computers and Society
Makes smart AI safer from bad uses.
SafeMLRM: Demystifying Safety in Multi-modal Large Reasoning Models
Machine Learning (CS)
Makes smart AI safer from bad instructions.
Red Teaming Large Reasoning Models
Cryptography and Security
Tests smart computers for honesty and safety.