Keep It Real: Challenges in Attacking Compression-Based Adversarial Purification
By: Samuel Räber , Till Aczel , Andreas Plesner and more
Potential Business Impact:
Makes AI images harder for hackers to trick.
Previous work has suggested that preprocessing images through lossy compression can defend against adversarial perturbations, but comprehensive attack evaluations have been lacking. In this paper, we construct strong white-box and adaptive attacks against various compression models and identify a critical challenge for attackers: high realism in reconstructed images significantly increases attack difficulty. Through rigorous evaluation across multiple attack scenarios, we demonstrate that compression models capable of producing realistic, high-fidelity reconstructions are substantially more resistant to our attacks. In contrast, low-realism compression models can be broken. Our analysis reveals that this is not due to gradient masking. Rather, realistic reconstructions maintaining distributional alignment with natural images seem to offer inherent robustness. This work highlights a significant obstacle for future adversarial attacks and suggests that developing more effective techniques to overcome realism represents an essential challenge for comprehensive security evaluation.
Similar Papers
Human Aligned Compression for Robust Models
CV and Pattern Recognition
Protects AI from fake pictures by cleaning them.
The Rate-Distortion-Perception Trade-Off with Algorithmic Realism
Information Theory
Makes pictures look real when shrunk down.
Model Compression vs. Adversarial Robustness: An Empirical Study on Language Models for Code
Software Engineering
Makes AI code checkers less safe when smaller.