Assessing the Noise Robustness of Class Activation Maps: A Framework for Reliable Model Interpretability
By: Syamantak Sarkar , Revoti P. Bora , Bhupender Kaushal and more
Potential Business Impact:
Makes AI see what's important, even with bad pictures.
Class Activation Maps (CAMs) are one of the important methods for visualizing regions used by deep learning models. Yet their robustness to different noise remains underexplored. In this work, we evaluate and report the resilience of various CAM methods for different noise perturbations across multiple architectures and datasets. By analyzing the influence of different noise types on CAM explanations, we assess the susceptibility to noise and the extent to which dataset characteristics may impact explanation stability. The findings highlight considerable variability in noise sensitivity for various CAMs. We propose a robustness metric for CAMs that captures two key properties: consistency and responsiveness. Consistency reflects the ability of CAMs to remain stable under input perturbations that do not alter the predicted class, while responsiveness measures the sensitivity of CAMs to changes in the prediction caused by such perturbations. The metric is evaluated empirically across models, different perturbations, and datasets along with complementary statistical tests to exemplify the applicability of our proposed approach.
Similar Papers
Region-CAM: Towards Accurate Object Regions in Class Activation Maps for Weakly Supervised Learning Tasks
CV and Pattern Recognition
Helps computers see the whole object better.
CF-CAM: Cluster Filter Class Activation Mapping for Reliable Gradient-Based Interpretability
Machine Learning (CS)
Shows how AI makes decisions, faster and better.
Metric-Guided Synthesis of Class Activation Mapping
CV and Pattern Recognition
Shows computers which parts of a picture matter.