Are Synthetic Corruptions A Reliable Proxy For Real-World Corruptions?
By: Shashank Agnihotri , David Schader , Nico Sharei and more
Potential Business Impact:
Tests computer vision with fake weather, not real.
Deep learning (DL) models are widely used in real-world applications but remain vulnerable to distribution shifts, especially due to weather and lighting changes. Collecting diverse real-world data for testing the robustness of DL models is resource-intensive, making synthetic corruptions an attractive alternative for robustness testing. However, are synthetic corruptions a reliable proxy for real-world corruptions? To answer this, we conduct the largest benchmarking study on semantic segmentation models, comparing performance on real-world corruptions and synthetic corruptions datasets. Our results reveal a strong correlation in mean performance, supporting the use of synthetic corruptions for robustness evaluation. We further analyze corruption-specific correlations, providing key insights to understand when synthetic corruptions succeed in representing real-world corruptions. Open-source Code: https://github.com/shashankskagnihotri/benchmarking_robustness/tree/segmentation_david/semantic_segmentation
Similar Papers
Benchmarking the Spatial Robustness of DNNs via Natural and Adversarial Localized Corruptions
CV and Pattern Recognition
Makes self-driving cars see better in bad weather.
DispBench: Benchmarking Disparity Estimation to Synthetic Corruptions
CV and Pattern Recognition
Tests how well AI sees depth in pictures.
Analysing the Robustness of Vision-Language-Models to Common Corruptions
CV and Pattern Recognition
Makes AI understand pictures even when they're messy.