An h-space Based Adversarial Attack for Protection Against Few-shot Personalization
By: Xide Xu , Sandesh Kamath , Muhammad Atif Butt and more
Potential Business Impact:
Stops AI from changing your private pictures.
The versatility of diffusion models in generating customized images from few samples raises significant privacy concerns, particularly regarding unauthorized modifications of private content. This concerning issue has renewed the efforts in developing protection mechanisms based on adversarial attacks, which generate effective perturbations to poison diffusion models. Our work is motivated by the observation that these models exhibit a high degree of abstraction within their semantic latent space (`h-space'), which encodes critical high-level features for generating coherent and meaningful content. In this paper, we propose a novel anti-customization approach, called HAAD (h-space based Adversarial Attack for Diffusion models), that leverages adversarial attacks to craft perturbations based on the h-space that can efficiently degrade the image generation process. Building upon HAAD, we further introduce a more efficient variant, HAAD-KV, that constructs perturbations solely based on the KV parameters of the h-space. This strategy offers a stronger protection, that is computationally less expensive. Despite their simplicity, our methods outperform state-of-the-art adversarial attacks, highlighting their effectiveness.
Similar Papers
Make the Most of Everything: Further Considerations on Disrupting Diffusion-based Customization
CV and Pattern Recognition
Stops AI from making fake pictures of you.
A Knowledge-guided Adversarial Defense for Resisting Malicious Visual Manipulation
CV and Pattern Recognition
Stops fake images from fooling people.
Latent Diffusion Unlearning: Protecting Against Unauthorized Personalization Through Trajectory Shifted Perturbations
CV and Pattern Recognition
Makes AI art harder to copy or steal.