Secure Human Oversight of AI: Exploring the Attack Surface of Human Oversight
By: Jonas C. Ditz , Veronika Lazar , Elmar Lichtmeß and more
Potential Business Impact:
Secures AI by protecting people watching it.
Human oversight of AI is promoted as a safeguard against risks such as inaccurate outputs, system malfunctions, or violations of fundamental rights, and is mandated in regulation like the European AI Act. Yet debates on human oversight have largely focused on its effectiveness, while overlooking a critical dimension: the security of human oversight. We argue that human oversight creates a new attack surface within the safety, security, and accountability architecture of AI operations. Drawing on cybersecurity perspectives, we analyze attack vectors that threaten the requirements of effective human oversight, thereby undermining the safety of AI operations. Such attacks may target the AI system, its communication with oversight personnel, or the personnel themselves. We then outline hardening strategies to mitigate these risks. Our contributions are: (1) introducing a security perspective on human oversight, and (2) providing an overview of attack vectors and hardening strategies to enable secure human oversight of AI.
Similar Papers
AI and Human Oversight: A Risk-Based Framework for Alignment
Computers and Society
Keeps AI from making bad choices without people.
Human-AI Complementarity: A Goal for Amplified Oversight
Artificial Intelligence
AI helps people check if AI is telling the truth.
Surveying the Operational Cybersecurity and Supply Chain Threat Landscape when Developing and Deploying AI Systems
Cryptography and Security
Protects AI from new kinds of computer attacks.