Score: 0

Narrowing Class-Wise Robustness Gaps in Adversarial Training

Published: March 20, 2025 | arXiv ID: 2503.16179v1

By: Fatemeh Amerehi, Patrick Healy

Potential Business Impact:

Makes AI better at guessing, even with tricky data.

Business Areas:
A/B Testing Data and Analytics

Efforts to address declining accuracy as a result of data shifts often involve various data-augmentation strategies. Adversarial training is one such method, designed to improve robustness to worst-case distribution shifts caused by adversarial examples. While this method can improve robustness, it may also hinder generalization to clean examples and exacerbate performance imbalances across different classes. This paper explores the impact of adversarial training on both overall and class-specific performance, as well as its spill-over effects. We observe that enhanced labeling during training boosts adversarial robustness by 53.50% and mitigates class imbalances by 5.73%, leading to improved accuracy in both clean and adversarial settings compared to standard adversarial training.

Country of Origin
🇮🇪 Ireland

Page Count
7 pages

Category
Computer Science:
CV and Pattern Recognition