Score: 1

Adversarial Surrogate Risk Bounds for Binary Classification

Published: June 11, 2025 | arXiv ID: 2506.09348v1

By: Natalie S. Frank

BigTech Affiliations: University of Washington

Potential Business Impact:

Makes AI harder for hackers to trick.

Business Areas:
Predictive Analytics Artificial Intelligence, Data and Analytics, Software

A central concern in classification is the vulnerability of machine learning models to adversarial attacks. Adversarial training is one of the most popular techniques for training robust classifiers, which involves minimizing an adversarial surrogate risk. Recent work characterized when a minimizing sequence of an adversarial surrogate risk is also a minimizing sequence of the adversarial classification risk for binary classification -- a property known as adversarial consistency. However, these results do not address the rate at which the adversarial classification risk converges to its optimal value for such a sequence of functions that minimize the adversarial surrogate. This paper provides surrogate risk bounds that quantify that convergence rate. Additionally, we derive distribution-dependent surrogate risk bounds in the standard (non-adversarial) learning setting, that may be of independent interest.

Country of Origin
🇺🇸 United States

Page Count
37 pages

Category
Computer Science:
Machine Learning (CS)