Adversarially Robust AI-Generated Image Detection for Free: An Information Theoretic Perspective
By: Ruixuan Zhang , He Wang , Zhengyu Zhao and more
Potential Business Impact:
Finds fake AI pictures even when tricked.
Rapid advances in Artificial Intelligence Generated Images (AIGI) have facilitated malicious use, such as forgery and misinformation. Therefore, numerous methods have been proposed to detect fake images. Although such detectors have been proven to be universally vulnerable to adversarial attacks, defenses in this field are scarce. In this paper, we first identify that adversarial training (AT), widely regarded as the most effective defense, suffers from performance collapse in AIGI detection. Through an information-theoretic lens, we further attribute the cause of collapse to feature entanglement, which disrupts the preservation of feature-label mutual information. Instead, standard detectors show clear feature separation. Motivated by this difference, we propose Training-free Robust Detection via Information-theoretic Measures (TRIM), the first training-free adversarial defense for AIGI detection. TRIM builds on standard detectors and quantifies feature shifts using prediction entropy and KL divergence. Extensive experiments across multiple datasets and attacks validate the superiority of our TRIM, e.g., outperforming the state-of-the-art defense by 33.88% (28.91%) on ProGAN (GenImage), while well maintaining original accuracy.
Similar Papers
Is Artificial Intelligence Generated Image Detection a Solved Problem?
CV and Pattern Recognition
Finds fake pictures made by computers.
Transferable Dual-Domain Feature Importance Attack against AI-Generated Image Detector
CV and Pattern Recognition
Tricks AI image detectors to see fake pictures.
Robustness in AI-Generated Detection: Enhancing Resistance to Adversarial Attacks
CV and Pattern Recognition
Stops fake faces from fooling computer detectors.