SFIBA: Spatial-based Full-target Invisible Backdoor Attacks
By: Yangxu Yin , Honglong Chen , Yudong Gao and more
Potential Business Impact:
Hides secret messages in pictures to trick computers.
Multi-target backdoor attacks pose significant security threats to deep neural networks, as they can preset multiple target classes through a single backdoor injection. This allows attackers to control the model to misclassify poisoned samples with triggers into any desired target class during inference, exhibiting superior attack performance compared with conventional backdoor attacks. However, existing multi-target backdoor attacks fail to guarantee trigger specificity and stealthiness in black-box settings, resulting in two main issues. First, they are unable to simultaneously target all classes when only training data can be manipulated, limiting their effectiveness in realistic attack scenarios. Second, the triggers often lack visual imperceptibility, making poisoned samples easy to detect. To address these problems, we propose a Spatial-based Full-target Invisible Backdoor Attack, called SFIBA. It restricts triggers for different classes to specific local spatial regions and morphologies in the pixel space to ensure specificity, while employing a frequency-domain-based trigger injection method to guarantee stealthiness. Specifically, for injection of each trigger, we first apply fast fourier transform to obtain the amplitude spectrum of clean samples in local spatial regions. Then, we employ discrete wavelet transform to extract the features from the amplitude spectrum and use singular value decomposition to integrate the trigger. Subsequently, we selectively filter parts of the trigger in pixel space to implement trigger morphology constraints and adjust injection coefficients based on visual effects. We conduct experiments on multiple datasets and models. The results demonstrate that SFIBA can achieve excellent attack performance and stealthiness, while preserving the model's performance on benign samples, and can also bypass existing backdoor defenses.
Similar Papers
FFCBA: Feature-based Full-target Clean-label Backdoor Attacks
Cryptography and Security
Makes AI models easily tricked by hidden patterns.
IPBA: Imperceptible Perturbation Backdoor Attack in Federated Self-Supervised Learning
Cryptography and Security
Makes AI models secretly learn wrong things.
CIS-BA: Continuous Interaction Space Based Backdoor Attack for Object Detection in the Real-World
CV and Pattern Recognition
Tricks self-driving cars to see wrong things.