Score: 2

Manipulating LLM Web Agents with Indirect Prompt Injection Attack via HTML Accessibility Tree

Published: July 20, 2025 | arXiv ID: 2507.14799v1

By: Sam Johnson, Viet Pham, Thai Le

Potential Business Impact:

Hackers can trick web robots into doing bad things.

Plain English Summary

Imagine AI assistants that can browse the internet for you, like booking appointments or finding information. This research found a sneaky way to trick these AI assistants into doing bad things, like stealing your passwords or making you click on ads, by hiding special code on websites. This is a big deal because as more companies use these AI assistants, we need to make sure they're safe from these kinds of attacks so our personal information stays protected.

This work demonstrates that LLM-based web navigation agents offer powerful automation capabilities but are vulnerable to Indirect Prompt Injection (IPI) attacks. We show that adversaries can embed universal adversarial triggers in webpage HTML to hijack agent behavior that utilizes the accessibility tree to parse HTML, causing unintended or malicious actions. Using the Greedy Coordinate Gradient (GCG) algorithm and a Browser Gym agent powered by Llama-3.1, our system demonstrates high success rates across real websites in both targeted and general attacks, including login credential exfiltration and forced ad clicks. Our empirical results highlight critical security risks and the need for stronger defenses as LLM-driven autonomous web agents become more widely adopted. The system software (https://github.com/sej2020/manipulating-web-agents) is released under the MIT License, with an accompanying publicly available demo website (http://lethaiq.github.io/attack-web-llm-agent).

Country of Origin
πŸ‡»πŸ‡³ πŸ‡ΊπŸ‡Έ Viet Nam, United States

Repos / Data Links

Page Count
10 pages

Category
Computer Science:
Cryptography and Security