ImageNet-trained CNNs are not biased towards texture: Revisiting feature reliance through controlled suppression
By: Tom Burgert , Oliver Stoll , Paolo Rota and more
Potential Business Impact:
Computers see shapes, not just textures.
The hypothesis that Convolutional Neural Networks (CNNs) are inherently texture-biased has shaped much of the discourse on feature use in deep learning. We revisit this hypothesis by examining limitations in the cue-conflict experiment by Geirhos et al. To address these limitations, we propose a domain-agnostic framework that quantifies feature reliance through systematic suppression of shape, texture, and color cues, avoiding the confounds of forced-choice conflicts. By evaluating humans and neural networks under controlled suppression conditions, we find that CNNs are not inherently texture-biased but predominantly rely on local shape features. Nonetheless, this reliance can be substantially mitigated through modern training strategies or architectures (ConvNeXt, ViTs). We further extend the analysis across computer vision, medical imaging, and remote sensing, revealing that reliance patterns differ systematically: computer vision models prioritize shape, medical imaging models emphasize color, and remote sensing models exhibit a stronger reliance towards texture. Code is available at https://github.com/tomburgert/feature-reliance.
Similar Papers
Promoting Shape Bias in CNNs: Frequency-Based and Contrastive Regularization for Corruption Robustness
CV and Pattern Recognition
Makes computers see objects even when they're blurry.
On the Relationship Between Double Descent of CNNs and Shape/Texture Bias Under Learning Process
CV and Pattern Recognition
Helps computers see better by understanding shapes and textures.
Shape Bias and Robustness Evaluation via Cue Decomposition for Image Classification and Segmentation
CV and Pattern Recognition
Helps AI understand pictures better, not just textures.