MCP-ITP: An Automated Framework for Implicit Tool Poisoning in MCP
By: Ruiqi Li , Zhiqiang Wang , Yunhao Yao and more
Potential Business Impact:
Hides bad computer instructions in tool descriptions.
To standardize interactions between LLM-based agents and their environments, the Model Context Protocol (MCP) was proposed and has since been widely adopted. However, integrating external tools expands the attack surface, exposing agents to tool poisoning attacks. In such attacks, malicious instructions embedded in tool metadata are injected into the agent context during MCP registration phase, thereby manipulating agent behavior. Prior work primarily focuses on explicit tool poisoning or relied on manually crafted poisoned tools. In contrast, we focus on a particularly stealthy variant: implicit tool poisoning, where the poisoned tool itself remains uninvoked. Instead, the instructions embedded in the tool metadata induce the agent to invoke a legitimate but high-privilege tool to perform malicious operations. We propose MCP-ITP, the first automated and adaptive framework for implicit tool poisoning within the MCP ecosystem. MCP-ITP formulates poisoned tool generation as a black-box optimization problem and employs an iterative optimization strategy that leverages feedback from both an evaluation LLM and a detection LLM to maximize Attack Success Rate (ASR) while evading current detection mechanisms. Experimental results on the MCPTox dataset across 12 LLM agents demonstrate that MCP-ITP consistently outperforms the manually crafted baseline, achieving up to 84.2% ASR while suppressing the Malicious Tool Detection Rate (MDR) to as low as 0.3%.
Similar Papers
MCPTox: A Benchmark for Tool Poisoning Attack on Real-World MCP Servers
Cryptography and Security
Protects smart AI from bad tool instructions.
Systematic Analysis of MCP Security
Cryptography and Security
Finds ways AI can be tricked by tools.
Securing the Model Context Protocol: Defending LLMs Against Tool Poisoning and Adversarial Attacks
Cryptography and Security
Secures AI tools from hidden, dangerous instructions.