Toward Safe, Trustworthy and Realistic Augmented Reality User Experience
By: Yanming Xiu
Potential Business Impact:
Keeps augmented reality safe from bad virtual things.
As augmented reality (AR) becomes increasingly integrated into everyday life, ensuring the safety and trustworthiness of its virtual content is critical. Our research addresses the risks of task-detrimental AR content, particularly that which obstructs critical information or subtly manipulates user perception. We developed two systems, ViDDAR and VIM-Sense, to detect such attacks using vision-language models (VLMs) and multimodal reasoning modules. Building on this foundation, we propose three future directions: automated, perceptually aligned quality assessment of virtual content; detection of multimodal attacks; and adaptation of VLMs for efficient and user-centered deployment on AR devices. Overall, our work aims to establish a scalable, human-aligned framework for safeguarding AR experiences and seeks feedback on perceptual modeling, multimodal AR content implementation, and lightweight model adaptation.
Similar Papers
Demonstrating Visual Information Manipulation Attacks in Augmented Reality: A Hands-On Miniature City-Based Setup
Human-Computer Interaction
Protects augmented reality from tricking your eyes.
A Neurosymbolic Framework for Interpretable Cognitive Attack Detection in Augmented Reality
CV and Pattern Recognition
Stops fake things in your AR from tricking you.
ARMs: Adaptive Red-Teaming Agent against Multimodal Models with Plug-and-Play Attacks
Artificial Intelligence
Finds hidden dangers in AI that sees and talks.