International AI Safety Report 2025: Second Key Update: Technical Safeguards and Risk Management
By: Yoshua Bengio , Stephen Clare , Carina Prunkl and more
Potential Business Impact:
Makes AI safer from being used for bad things.
This second update to the 2025 International AI Safety Report assesses new developments in general-purpose AI risk management over the past year. It examines how researchers, public institutions, and AI developers are approaching risk management for general-purpose AI. In recent months, for example, three leading AI developers applied enhanced safeguards to their new models, as their internal pre-deployment testing could not rule out the possibility that these models could be misused to help create biological weapons. Beyond specific precautionary measures, there have been a range of other advances in techniques for making AI models and systems more reliable and resistant to misuse. These include new approaches in adversarial training, data curation, and monitoring systems. In parallel, institutional frameworks that operationalise and formalise these technical capabilities are starting to emerge: the number of companies publishing Frontier AI Safety Frameworks more than doubled in 2025, and governments and international organisations have established a small number of governance frameworks for general-purpose AI, focusing largely on transparency and risk assessment.
Similar Papers
International AI Safety Report 2025: First Key Update: Capabilities and Risk Implications
Computers and Society
AI solves harder problems, but still makes mistakes.
AI Safety is Stuck in Technical Terms -- A System Safety Response to the International AI Safety Report
Computers and Society
Makes AI safer by looking at all its parts.
Evaluating AI Companies' Frontier Safety Frameworks: Methodology and Results
Computers and Society
Helps AI companies build safer, more responsible systems.