A Knowledge-guided Adversarial Defense for Resisting Malicious Visual Manipulation
By: Dawei Zhou , Suzhi Gang , Decheng Liu and more
Potential Business Impact:
Stops fake images from fooling people.
Malicious applications of visual manipulation have raised serious threats to the security and reputation of users in many fields. To alleviate these issues, adversarial noise-based defenses have been enthusiastically studied in recent years. However, ``data-only" methods tend to distort fake samples in the low-level feature space rather than the high-level semantic space, leading to limitations in resisting malicious manipulation. Frontier research has shown that integrating knowledge in deep learning can produce reliable and generalizable solutions. Inspired by these, we propose a knowledge-guided adversarial defense (KGAD) to actively force malicious manipulation models to output semantically confusing samples. Specifically, in the process of generating adversarial noise, we focus on constructing significant semantic confusions at the domain-specific knowledge level, and exploit a metric closely related to visual perception to replace the general pixel-wise metrics. The generated adversarial noise can actively interfere with the malicious manipulation model by triggering knowledge-guided and perception-related disruptions in the fake samples. To validate the effectiveness of the proposed method, we conduct qualitative and quantitative experiments on human perception and visual quality assessment. The results on two different tasks both show that our defense provides better protection compared to state-of-the-art methods and achieves great generalizability.
Similar Papers
Adversarial-Guided Diffusion for Multimodal LLM Attacks
CV and Pattern Recognition
Tricks AI into seeing fake things in pictures.
Dual Attention Guided Defense Against Malicious Edits
CV and Pattern Recognition
Stops AI from making fake pictures from words.
An h-space Based Adversarial Attack for Protection Against Few-shot Personalization
CV and Pattern Recognition
Stops AI from changing your private pictures.