Securing AI Agent Execution
By: Christoph Bühler , Matteo Biagiola , Luca Di Grazia and more
Potential Business Impact:
Protects AI tools from hackers.
Large Language Models (LLMs) have evolved into AI agents that interact with external tools and environments to perform complex tasks. The Model Context Protocol (MCP) has become the de facto standard for connecting agents with such resources, but security has lagged behind: thousands of MCP servers execute with unrestricted access to host systems, creating a broad attack surface. In this paper, we introduce AgentBound, the first access control framework for MCP servers. AgentBound combines a declarative policy mechanism, inspired by the Android permission model, with a policy enforcement engine that contains malicious behavior without requiring MCP server modifications. We build a dataset containing the 296 most popular MCP servers, and show that access control policies can be generated automatically from source code with 80.9% accuracy. We also show that AgentBound blocks the majority of security threats in several malicious MCP servers, and that policy enforcement engine introduces negligible overhead. Our contributions provide developers and project managers with a practical foundation for securing MCP servers while maintaining productivity, enabling researchers and tool builders to explore new directions for declarative access control and MCP security.
Similar Papers
Securing AI Agent Execution
Cryptography and Security
Protects AI tools from hackers and bad code.
Securing the Model Context Protocol (MCP): Risks, Controls, and Governance
Cryptography and Security
Secures AI agents from hackers and mistakes.
MCPGuard : Automatically Detecting Vulnerabilities in MCP Servers
Cryptography and Security
Fixes security holes in smart AI tools.