GUARDIAN: Safeguarding LLM Multi-Agent Collaborations with Temporal Graph Modeling
By: Jialong Zhou, Lichao Wang, Xiao Yang
Potential Business Impact:
Keeps AI helpers from making big mistakes.
The emergence of large language models (LLMs) enables the development of intelligent agents capable of engaging in complex and multi-turn dialogues. However, multi-agent collaboration face critical safety challenges, such as hallucination amplification and error injection and propagation. This paper presents GUARDIAN, a unified method for detecting and mitigating multiple safety concerns in GUARDing Intelligent Agent collaboratioNs. By modeling the multi-agent collaboration process as a discrete-time temporal attributed graph, GUARDIAN explicitly captures the propagation dynamics of hallucinations and errors. The unsupervised encoder-decoder architecture incorporating an incremental training paradigm, learns to reconstruct node attributes and graph structures from latent embeddings, enabling the identification of anomalous nodes and edges with unparalleled precision. Moreover, we introduce a graph abstraction mechanism based on the Information Bottleneck Theory, which compresses temporal interaction graphs while preserving essential patterns. Extensive experiments demonstrate GUARDIAN's effectiveness in safeguarding LLM multi-agent collaborations against diverse safety vulnerabilities, achieving state-of-the-art accuracy with efficient resource utilization.
Similar Papers
LogicGuard: Improving Embodied LLM agents through Temporal Logic based Critics
Artificial Intelligence
Makes AI robots follow rules and finish tasks.
G-Safeguard: A Topology-Guided Security Lens and Treatment on LLM-based Multi-agent Systems
Cryptography and Security
Protects smart AI teams from being tricked.
DialogGuard: Multi-Agent Psychosocial Safety Evaluation of Sensitive LLM Responses
Artificial Intelligence
Tests AI for safe and helpful online chats.