Uncovering Anomalous Events for Marine Environmental Monitoring via Visual Anomaly Detection
By: Laura Weihl , Nejc Novak , Stefan H. Bengtson and more
Potential Business Impact:
Finds rare sea creatures in hours of video.
Underwater video monitoring is a promising strategy for assessing marine biodiversity, but the vast volume of uneventful footage makes manual inspection highly impractical. In this work, we explore the use of visual anomaly detection (VAD) based on deep neural networks to automatically identify interesting or anomalous events. We introduce AURA, the first multi-annotator benchmark dataset for underwater VAD, and evaluate four VAD models across two marine scenes. We demonstrate the importance of robust frame selection strategies to extract meaningful video segments. Our comparison against multiple annotators reveals that VAD performance of current models varies dramatically and is highly sensitive to both the amount of training data and the variability in visual content that defines "normal" scenes. Our results highlight the value of soft and consensus labels and offer a practical approach for supporting scientific exploration and scalable biodiversity monitoring.
Similar Papers
A Survey on Video Anomaly Detection via Deep Learning: Human, Vehicle, and Environment
CV and Pattern Recognition
Finds weird things happening in videos.
ALFred: An Active Learning Framework for Real-world Semi-supervised Anomaly Detection with Adaptive Thresholds
CV and Pattern Recognition
Spots weird things in videos, even when things change.
DUAL-VAD: Dual Benchmarks and Anomaly-Focused Sampling for Video Anomaly Detection
CV and Pattern Recognition
Finds weird things happening in videos.