International Agreements on AI Safety: Review and Recommendations for a Conditional AI Safety Treaty
By: Rebecca Scholefield, Samuel Martin, Otto Barten
Potential Business Impact:
Makes AI development safer by setting rules.
The malicious use or malfunction of advanced general-purpose AI (GPAI) poses risks that, according to leading experts, could lead to the 'marginalisation or extinction of humanity.' To address these risks, there are an increasing number of proposals for international agreements on AI safety. In this paper, we review recent (2023-) proposals, identifying areas of consensus and disagreement, and drawing on related literature to assess their feasibility. We focus our discussion on risk thresholds, regulations, types of international agreement and five related processes: building scientific consensus, standardisation, auditing, verification and incentivisation. Based on this review, we propose a treaty establishing a compute threshold above which development requires rigorous oversight. This treaty would mandate complementary audits of models, information security and governance practices, overseen by an international network of AI Safety Institutes (AISIs) with authority to pause development if risks are unacceptable. Our approach combines immediately implementable measures with a flexible structure that can adapt to ongoing research.
Similar Papers
An International Agreement to Prevent the Premature Creation of Artificial Superintelligence
Computers and Society
Stops super-smart computers from being built too soon.
International AI Safety Report
Computers and Society
Makes AI safer for everyone to use.
International AI Safety Report 2025: Second Key Update: Technical Safeguards and Risk Management
Computers and Society
Makes AI safer from being used for bad things.