A Statistical Method for Attack-Agnostic Adversarial Attack Detection with Compressive Sensing Comparison
By: Chinthana Wimalasuriya, Spyros Tragoudas
Potential Business Impact:
Stops sneaky computer tricks from fooling smart programs.
Adversarial attacks present a significant threat to modern machine learning systems. Yet, existing detection methods often lack the ability to detect unseen attacks or detect different attack types with a high level of accuracy. In this work, we propose a statistical approach that establishes a detection baseline before a neural network's deployment, enabling effective real-time adversarial detection. We generate a metric of adversarial presence by comparing the behavior of a compressed/uncompressed neural network pair. Our method has been tested against state-of-the-art techniques, and it achieves near-perfect detection across a wide range of attack types. Moreover, it significantly reduces false positives, making it both reliable and practical for real-world applications.
Similar Papers
Adversarially-Aware Architecture Design for Robust Medical AI Systems
Machine Learning (CS)
Protects AI from tricks that harm patients.
A unified Bayesian framework for adversarial robustness
Machine Learning (Stat)
Protects computer brains from sneaky tricks.
Beyond Vulnerabilities: A Survey of Adversarial Attacks as Both Threats and Defenses in Computer Vision Systems
CV and Pattern Recognition
Strengthens AI against sneaky image tricks