Labels Matter More Than Models: Quantifying the Benefit of Supervised Time Series Anomaly Detection
By: Zhijie Zhong , Zhiwen Yu , Kaixiang Yang and more
Potential Business Impact:
Find problems in data using a few examples.
Time series anomaly detection (TSAD) is a critical data mining task often constrained by label scarcity. Consequently, current research predominantly focuses on Unsupervised Time-series Anomaly Detection (UTAD), relying on complex architectures to model normal data distributions. However, this approach often overlooks the significant performance gains available from limited anomaly labels achievable in practical scenarios. This paper challenges the premise that architectural complexity is the optimal path for TSAD. We conduct the first methodical comparison between supervised and unsupervised paradigms and introduce STAND, a streamlined supervised baseline. Extensive experiments on five public datasets demonstrate that: (1) Labels matter more than models: under a limited labeling budget, simple supervised models significantly outperform complex state-of-the-art unsupervised methods; (2) Supervision yields higher returns: the performance gain from minimal supervision far exceeds that from architectural innovations; and (3) Practicality: STAND exhibits superior prediction consistency and anomaly localization compared to unsupervised counterparts. These findings advocate for a data-centric shift in TSAD research, emphasizing label utilization over purely algorithmic complexity. The code is publicly available at https://github.com/EmorZz1G/STAND.
Similar Papers
Harnessing Vision-Language Models for Time Series Anomaly Detection
CV and Pattern Recognition
Find weird patterns in data using AI.
VISTA: Unsupervised 2D Temporal Dependency Representations for Time Series Anomaly Detection
Machine Learning (CS)
Finds hidden problems in data without needing training.
TShape: Rescuing Machine Learning Models from Complex Shapelet Anomalies
Software Engineering
Finds weird patterns in computer data faster.