Score: 0

On the existence of consistent adversarial attacks in high-dimensional linear classification

Published: June 14, 2025 | arXiv ID: 2506.12454v1

By: Matteo Vilucchio, Lenka Zdeborová, Bruno Loureiro

Potential Business Impact:

Finds how computer mistakes can be tricked.

Business Areas:
A/B Testing Data and Analytics

What fundamentally distinguishes an adversarial attack from a misclassification due to limited model expressivity or finite data? In this work, we investigate this question in the setting of high-dimensional binary classification, where statistical effects due to limited data availability play a central role. We introduce a new error metric that precisely capture this distinction, quantifying model vulnerability to consistent adversarial attacks -- perturbations that preserve the ground-truth labels. Our main technical contribution is an exact and rigorous asymptotic characterization of these metrics in both well-specified models and latent space models, revealing different vulnerability patterns compared to standard robust error measures. The theoretical results demonstrate that as models become more overparameterized, their vulnerability to label-preserving perturbations grows, offering theoretical insight into the mechanisms underlying model sensitivity to adversarial attacks.

Page Count
33 pages

Category
Statistics:
Machine Learning (Stat)