Comparing AI Agents to Cybersecurity Professionals in Real-World Penetration Testing
By: Justin W. Lin , Eliot Krzysztof Jones , Donovan Julian Jasper and more
We present the first comprehensive evaluation of AI agents against human cybersecurity professionals in a live enterprise environment. We evaluate ten cybersecurity professionals alongside six existing AI agents and ARTEMIS, our new agent scaffold, on a large university network consisting of ~8,000 hosts across 12 subnets. ARTEMIS is a multi-agent framework featuring dynamic prompt generation, arbitrary sub-agents, and automatic vulnerability triaging. In our comparative study, ARTEMIS placed second overall, discovering 9 valid vulnerabilities with an 82% valid submission rate and outperforming 9 of 10 human participants. While existing scaffolds such as Codex and CyAgent underperformed relative to most human participants, ARTEMIS demonstrated technical sophistication and submission quality comparable to the strongest participants. We observe that AI agents offer advantages in systematic enumeration, parallel exploitation, and cost -- certain ARTEMIS variants cost $18/hour versus $60/hour for professional penetration testers. We also identify key capability gaps: AI agents exhibit higher false-positive rates and struggle with GUI-based tasks.
Similar Papers
Cybersecurity AI: Evaluating Agentic Cybersecurity in Attack/Defense CTFs
Cryptography and Security
AI is better at defending computers than attacking.
Shell or Nothing: Real-World Benchmarks and Memory-Activated Agents for Automated Penetration Testing
Cryptography and Security
AI finds computer weaknesses faster and cheaper.
Shell or Nothing: Real-World Benchmarks and Memory-Activated Agents for Automated Penetration Testing
Cryptography and Security
AI finds computer weaknesses faster and cheaper.