Empirical evaluation of the Frank-Wolfe methods for constructing white-box adversarial attacks
By: Kristina Korotkova, Aleksandr Katrutsa
Potential Business Impact:
Makes AI harder to trick with fake images.
The construction of adversarial attacks for neural networks appears to be a crucial challenge for their deployment in various services. To estimate the adversarial robustness of a neural network, a fast and efficient approach is needed to construct adversarial attacks. Since the formalization of adversarial attack construction involves solving a specific optimization problem, we consider the problem of constructing an efficient and effective adversarial attack from a numerical optimization perspective. Specifically, we suggest utilizing advanced projection-free methods, known as modified Frank-Wolfe methods, to construct white-box adversarial attacks on the given input data. We perform a theoretical and numerical evaluation of these methods and compare them with standard approaches based on projection operations or geometrical intuition. Numerical experiments are performed on the MNIST and CIFAR-10 datasets, utilizing a multiclass logistic regression model, the convolutional neural networks (CNNs), and the Vision Transformer (ViT).
Similar Papers
Adversarial Attacks Leverage Interference Between Features in Superposition
Machine Learning (CS)
Makes AI easier to trick by how it learns.
A Statistical Method for Attack-Agnostic Adversarial Attack Detection with Compressive Sensing Comparison
Cryptography and Security
Stops sneaky computer tricks from fooling smart programs.
Towards Trustworthy Wi-Fi Sensing: Systematic Evaluation of Deep Learning Model Robustness to Adversarial Attacks
Machine Learning (CS)
Makes wireless sensing safer from hacking.