Towards Real Unsupervised Anomaly Detection Via Confident Meta-Learning
By: Muhammad Aqeel , Shakiba Sharifi , Marco Cristani and more
Potential Business Impact:
Finds bad things even in messy data.
So-called unsupervised anomaly detection is better described as semi-supervised, as it assumes all training data are nominal. This assumption simplifies training but requires manual data curation, introducing bias and limiting adaptability. We propose Confident Meta-learning (CoMet), a novel training strategy that enables deep anomaly detection models to learn from uncurated datasets where nominal and anomalous samples coexist, eliminating the need for explicit filtering. Our approach integrates Soft Confident Learning, which assigns lower weights to low-confidence samples, and Meta-Learning, which stabilizes training by regularizing updates based on training validation loss covariance. This prevents overfitting and enhances robustness to noisy data. CoMet is model-agnostic and can be applied to any anomaly detection method trainable via gradient descent. Experiments on MVTec-AD, VIADUCT, and KSDD2 with two state-of-the-art models demonstrate the effectiveness of our approach, consistently improving over the baseline methods, remaining insensitive to anomalies in the training set, and setting a new state-of-the-art across all datasets.
Similar Papers
A Contrastive Learning-Guided Confident Meta-learning for Zero Shot Anomaly Detection
CV and Pattern Recognition
Finds weird things in data without examples.
Robust Anomaly Detection in Industrial Environments via Meta-Learning
CV and Pattern Recognition
Finds bad parts even with wrong labels.
Bridging Unsupervised and Semi-Supervised Anomaly Detection: A Theoretically-Grounded and Practical Framework with Synthetic Anomalies
Machine Learning (Stat)
Finds hidden problems by creating fake ones.