Language-Guided Open-World Anomaly Segmentation
By: Klara Reichard , Nikolas Brasch , Nassir Navab and more
Potential Business Impact:
Names unknown things for self-driving cars.
Open-world and anomaly segmentation methods seek to enable autonomous driving systems to detect and segment both known and unknown objects in real-world scenes. However, existing methods do not assign semantically meaningful labels to unknown regions, and distinguishing and learning representations for unknown classes remains difficult. While open-vocabulary segmentation methods show promise in generalizing to novel classes, they require a fixed inference vocabulary and thus cannot be directly applied to anomaly segmentation where unknown classes are unconstrained. We propose Clipomaly, the first CLIP-based open-world and anomaly segmentation method for autonomous driving. Our zero-shot approach requires no anomaly-specific training data and leverages CLIP's shared image-text embedding space to both segment unknown objects and assign human-interpretable names to them. Unlike open-vocabulary methods, our model dynamically extends its vocabulary at inference time without retraining, enabling robust detection and naming of anomalies beyond common class definitions such as those in Cityscapes. Clipomaly achieves state-of-the-art performance on established anomaly segmentation benchmarks while providing interpretability and flexibility essential for practical deployment.
Similar Papers
ClimaOoD: Improving Anomaly Segmentation via Physically Realistic Synthetic Data
CV and Pattern Recognition
Makes self-driving cars better at spotting unusual things.
Open Vocabulary Panoptic Segmentation With Retrieval Augmentation
CV and Pattern Recognition
Lets computers see any object, even new ones.
Leveraging Text-Driven Semantic Variation for Robust OOD Segmentation
CV and Pattern Recognition
Helps self-driving cars spot unexpected things.