Calibrated Adversarial Sampling: Multi-Armed Bandit-Guided Generalization Against Unforeseen Attacks
By: Rui Wang , Zeming Wei , Xiyue Zhang and more
Potential Business Impact:
Makes AI smarter and safer from tricks.
Deep Neural Networks (DNNs) are known to be vulnerable to various adversarial perturbations. To address the safety concerns arising from these vulnerabilities, adversarial training (AT) has emerged as one of the most effective paradigms for enhancing the robustness of DNNs. However, existing AT frameworks primarily focus on a single or a limited set of attack types, leaving DNNs still exposed to attack types that may be encountered in practice but not addressed during training. In this paper, we propose an efficient fine-tuning method called Calibrated Adversarial Sampling (CAS) to address these issues. From the optimization perspective within the multi-armed bandit framework, it dynamically designs rewards and balances exploration and exploitation by considering the dynamic and interdependent characteristics of multiple robustness dimensions. Experiments on benchmark datasets show that CAS achieves superior overall robustness while maintaining high clean accuracy, providing a new paradigm for robust generalization of DNNs.
Similar Papers
AdaGAT: Adaptive Guidance Adversarial Training for the Robustness of Deep Neural Networks
CV and Pattern Recognition
Makes small computer brains smarter and tougher.
Constrained Adversarial Perturbation
Machine Learning (CS)
Makes AI fooled by fake data less often.
DARD: Dice Adversarial Robustness Distillation against Adversarial Attacks
Machine Learning (CS)
Makes AI smarter and safer from tricks.