On the Escaping Efficiency of Distributed Adversarial Training Algorithms
By: Ying Cao, Kun Yuan, Ali H. Sayed
Potential Business Impact:
Makes AI stronger against tricky computer tricks.
Adversarial training has been widely studied in recent years due to its role in improving model robustness against adversarial attacks. This paper focuses on comparing different distributed adversarial training algorithms--including centralized and decentralized strategies--within multi-agent learning environments. Previous studies have highlighted the importance of model flatness in determining robustness. To this end, we develop a general theoretical framework to study the escaping efficiency of these algorithms from local minima, which is closely related to the flatness of the resulting models. We show that when the perturbation bound is sufficiently small (i.e., when the attack strength is relatively mild) and a large batch size is used, decentralized adversarial training algorithms--including consensus and diffusion--are guaranteed to escape faster from local minima than the centralized strategy, thereby favoring flatter minima. However, as the perturbation bound increases, this trend may no longer hold. In the simulation results, we illustrate our theoretical findings and systematically compare the performance of models obtained through decentralized and centralized adversarial training algorithms. The results highlight the potential of decentralized strategies to enhance the robustness of models in distributed settings.
Similar Papers
Stability and Generalization of Adversarial Diffusion Training
Machine Learning (CS)
Makes AI learn better even when tricked.
Fully-Decentralized MADDPG with Networked Agents
Machine Learning (CS)
Lets many AI agents learn to work together.
Distributional Adversarial Attacks and Training in Deep Hedging
Optimization and Control
Makes money-saving computer plans safer from market changes.