Aegis: Taxonomy and Optimizations for Overcoming Agent-Environment Failures in LLM Agents
By: Kevin Song , Anand Jayarajan , Yaoyao Ding and more
Potential Business Impact:
Helps smart computer programs finish jobs better.
Large Language Models (LLMs) agents augmented with domain tools promise to autonomously execute complex tasks requiring human-level intelligence, such as customer service and digital assistance. However, their practical deployment is often limited by their low success rates under complex real-world environments. To tackle this, prior research has primarily focused on improving the agents themselves, such as developing strong agentic LLMs, while overlooking the role of the system environment in which the agent operates. In this paper, we study a complementary direction: improving agent success rates by optimizing the system environment in which the agent operates. We collect 142 agent traces (3,656 turns of agent-environment interactions) across 5 state-of-the-art agentic benchmarks. By analyzing these agent failures, we propose a taxonomy for agent-environment interaction failures that includes 6 failure modes. Guided by these findings, we design Aegis, a set of targeted environment optimizations: 1) environment observability enhancement, 2) common computation offloading, and 3) speculative agentic actions. These techniques improve agent success rates on average by 6.7-12.5%, without any modifications to the agent and underlying LLM.
Similar Papers
AegisAgent: An Autonomous Defense Agent Against Prompt Injection Attacks in LLM-HARs
Cryptography and Security
Protects smart watches from sneaky instructions.
Aegis: Automated Error Generation and Identification for Multi-Agent Systems
Robotics
Finds mistakes in smart robot teams.
TraceAegis: Securing LLM-Based Agents via Hierarchical and Behavioral Anomaly Detection
Cryptography and Security
Protects smart computer helpers from being tricked.