Promoting Shape Bias in CNNs: Frequency-Based and Contrastive Regularization for Corruption Robustness
By: Robin Narsingh Ranabhat , Longwei Wang , Amit Kumar Patel and more
Potential Business Impact:
Makes computers see objects even when they're blurry.
Convolutional Neural Networks (CNNs) excel at image classification but remain vulnerable to common corruptions that humans handle with ease. A key reason for this fragility is their reliance on local texture cues rather than global object shapes -- a stark contrast to human perception. To address this, we propose two complementary regularization strategies designed to encourage shape-biased representations and enhance robustness. The first introduces an auxiliary loss that enforces feature consistency between original and low-frequency filtered inputs, discouraging dependence on high-frequency textures. The second incorporates supervised contrastive learning to structure the feature space around class-consistent, shape-relevant representations. Evaluated on the CIFAR-10-C benchmark, both methods improve corruption robustness without degrading clean accuracy. Our results suggest that loss-level regularization can effectively steer CNNs toward more shape-aware, resilient representations.
Similar Papers
ImageNet-trained CNNs are not biased towards texture: Revisiting feature reliance through controlled suppression
CV and Pattern Recognition
Computers see shapes, not just textures.
Learning Fourier shapes to probe the geometric world of deep neural networks
CV and Pattern Recognition
Teaches computers to see shapes, not just textures.
Mimicking Human Visual Development for Learning Robust Image Representations
CV and Pattern Recognition
Teaches computers to see better by starting blurry.