A Unified Bilevel Model for Adversarial Learning and A Case Study
By: Yutong Zheng, Qingna Li
Potential Business Impact:
Makes computer groups fooled by tiny data changes.
Adversarial learning has been attracting more and more attention thanks to the fast development of machine learning and artificial intelligence. However, due to the complicated structure of most machine learning models, the mechanism of adversarial attacks is not well interpreted. How to measure the effect of attack is still not quite clear. In this paper, we propose a unified bilevel model for adversarial learning. We further investigate the adversarial attack in clustering models and interpret it from data perturbation point of view. We reveal that when the data perturbation is relatively small, the clustering model is robust, whereas if it is relatively large, the clustering result changes, which leads to an attack. To measure the effect of attacks for clustering models, we analyse the well-definedness of the so-called $\delta$-measure, which can be used in the proposed bilevel model for adversarial learning of clustering models.
Similar Papers
A unified Bayesian framework for adversarial robustness
Machine Learning (Stat)
Protects computer brains from sneaky tricks.
Adversarially-Aware Architecture Design for Robust Medical AI Systems
Machine Learning (CS)
Protects AI from tricks that harm patients.
Selective Adversarial Attacks on LLM Benchmarks
Machine Learning (CS)
Makes AI tests unfair by tricking them.