Securing the Model Context Protocol (MCP): Risks, Controls, and Governance
By: Herman Errico, Jiquan Ngiam, Shanita Sojan
Potential Business Impact:
Secures AI agents from hackers and mistakes.
The Model Context Protocol (MCP) replaces static, developer-controlled API integrations with more dynamic, user-driven agent systems, which also introduces new security risks. As MCP adoption grows across community servers and major platforms, organizations encounter threats that existing AI governance frameworks (such as NIST AI RMF and ISO/IEC 42001) do not yet cover in detail. We focus on three types of adversaries that take advantage of MCP s flexibility: content-injection attackers that embed malicious instructions into otherwise legitimate data; supply-chain attackers who distribute compromised servers; and agents who become unintentional adversaries by over-stepping their role. Based on early incidents and proof-of-concept attacks, we describe how MCP can increase the attack surface through data-driven exfiltration, tool poisoning, and cross-system privilege escalation. In response, we propose a set of practical controls, including per-user authentication with scoped authorization, provenance tracking across agent workflows, containerized sandboxing with input/output checks, inline policy enforcement with DLP and anomaly detection, and centralized governance using private registries or gateway layers. The aim is to help organizations ensure that unvetted code does not run outside a sandbox, tools are not used beyond their intended scope, data exfiltration attempts are detectable, and actions can be audited end-to-end. We close by outlining open research questions around verifiable registries, formal methods for these dynamic systems, and privacy-preserving agent operations.
Similar Papers
MCPGuard : Automatically Detecting Vulnerabilities in MCP Servers
Cryptography and Security
Fixes security holes in smart AI tools.
Toward Understanding Security Issues in the Model Context Protocol Ecosystem
Cryptography and Security
Finds and fixes security flaws in AI tools.
Systematization of Knowledge: Security and Safety in the Model Context Protocol Ecosystem
Cryptography and Security
Secures AI agents from dangerous commands.