Toward a Safe Internet of Agents
By: Juan A. Wibowo, George C. Polyzos
Potential Business Impact:
Makes AI agents safer and more trustworthy.
Background: Autonomous agents powered by Large Language Models (LLMs) are driving a paradigm shift toward an "Internet of Agents" (IoA). While offering immense potential, this vision also introduces novel and systemic risks to safety and security. Objectives: Unlike common threat-centric taxonomies, our survey provides a principled, architectural framework for engineering safe and reliable agentic systems. We aim to identify the architectural sources of vulnerabilities to establish a foundation for secure design. Methods: We perform a bottom-up deconstruction of agentic systems, treating each component as a dual-use interface. The analysis spans three levels of complexity: the foundational Single Agent, the collaborative Multi-Agent System (MAS), and the visionary Interoperable Multi-Agent System (IMAS). At each level, we identify core architectural components and their inherent security risks. Results & Conclusions: Our central finding is that agentic safety is an architectural principle, not an add-on. By identifying specific vulnerabilities and deriving mitigation principles at each level of the agentic stack, this survey serves as a foundational guide for building the capable, safe, and trustworthy AI needed to realize a secure Internet of Agents.
Similar Papers
Security of Internet of Agents: Attacks and Countermeasures
Cryptography and Security
Keeps smart AI helpers safe and private online.
Formalizing the Safety, Security, and Functional Properties of Agentic AI Systems
Artificial Intelligence
Makes smart robots work together safely and reliably.
Decentralized Multi-Agent System with Trust-Aware Communication
Multiagent Systems
Builds safer, smarter robot teams that can't be stopped.