Non-Parametric Probabilistic Robustness: A Conservative Metric with Optimized Perturbation Distributions
By: Zheng Wang , Yi Zhang , Siddartha Khastgir and more
Potential Business Impact:
Makes AI more trustworthy with unknown errors.
Deep learning (DL) models, despite their remarkable success, remain vulnerable to small input perturbations that can cause erroneous outputs, motivating the recent proposal of probabilistic robustness (PR) as a complementary alternative to adversarial robustness (AR). However, existing PR formulations assume a fixed and known perturbation distribution, an unrealistic expectation in practice. To address this limitation, we propose non-parametric probabilistic robustness (NPPR), a more practical PR metric that does not rely on any predefined perturbation distribution. Following the non-parametric paradigm in statistical modeling, NPPR learns an optimized perturbation distribution directly from data, enabling conservative PR evaluation under distributional uncertainty. We further develop an NPPR estimator based on a Gaussian Mixture Model (GMM) with Multilayer Perceptron (MLP) heads and bicubic up-sampling, covering various input-dependent and input-independent perturbation scenarios. Theoretical analyses establish the relationships among AR, PR, and NPPR. Extensive experiments on CIFAR-10, CIFAR-100, and Tiny ImageNet across ResNet18/50, WideResNet50 and VGG16 validate NPPR as a more practical robustness metric, showing up to 40\% more conservative (lower) PR estimates compared to assuming those common perturbation distributions used in state-of-the-arts.
Similar Papers
Probabilistic Robustness for Free? Revisiting Training via a Benchmark
CV and Pattern Recognition
Tests computer smarts against tricky mistakes.
Get Global Guarantees: On the Probabilistic Nature of Perturbation Robustness
Machine Learning (CS)
Makes AI safer by testing its mistakes.
Enhancing Robustness of Implicit Neural Representations Against Weight Perturbations
CV and Pattern Recognition
Makes AI models harder to trick with bad data.