Adversarial generalization of unfolding (model-based) networks
By: Vicky Kouni
Potential Business Impact:
Makes AI safer from sneaky tricks.
Unfolding networks are interpretable networks emerging from iterative algorithms, incorporate prior knowledge of data structure, and are designed to solve inverse problems like compressed sensing, which deals with recovering data from noisy, missing observations. Compressed sensing finds applications in critical domains, from medical imaging to cryptography, where adversarial robustness is crucial to prevent catastrophic failures. However, a solid theoretical understanding of the performance of unfolding networks in the presence of adversarial attacks is still in its infancy. In this paper, we study the adversarial generalization of unfolding networks when perturbed with $l_2$-norm constrained attacks, generated by the fast gradient sign method. Particularly, we choose a family of state-of-the-art overaparameterized unfolding networks and deploy a new framework to estimate their adversarial Rademacher complexity. Given this estimate, we provide adversarial generalization error bounds for the networks under study, which are tight with respect to the attack level. To our knowledge, this is the first theoretical analysis on the adversarial generalization of unfolding networks. We further present a series of experiments on real-world data, with results corroborating our derived theory, consistently for all data. Finally, we observe that the family's overparameterization can be exploited to promote adversarial robustness, shedding light on how to efficiently robustify neural networks.
Similar Papers
Stability and Generalization of Adversarial Diffusion Training
Machine Learning (CS)
Makes AI learn better even when tricked.
Deep Unfolding: Recent Developments, Theory, and Design Guidelines
Machine Learning (CS)
Turns slow math problems into fast computer learning.
Machine Learning-based Unfolding for Cross Section Measurements in the Presence of Nuisance Parameters
Applications
Fixes science measurements distorted by machines.