Prototypical Contrastive Learning For Improved Few-Shot Audio Classification
By: Christos Sgouropoulos , Christos Nikou , Stefanos Vlachos and more
Potential Business Impact:
Teaches computers to learn sounds with few examples.
Few-shot learning has emerged as a powerful paradigm for training models with limited labeled data, addressing challenges in scenarios where large-scale annotation is impractical. While extensive research has been conducted in the image domain, few-shot learning in audio classification remains relatively underexplored. In this work, we investigate the effect of integrating supervised contrastive loss into prototypical few shot training for audio classification. In detail, we demonstrate that angular loss further improves the performance compared to the standard contrastive loss. Our method leverages SpecAugment followed by a self-attention mechanism to encapsulate diverse information of augmented input versions into one unified embedding. We evaluate our approach on MetaAudio, a benchmark including five datasets with predefined splits, standardized preprocessing, and a comprehensive set of few-shot learning models for comparison. The proposed approach achieves state-of-the-art performance in a 5-way, 5-shot setting.
Similar Papers
Improving Audio Classification by Transitioning from Zero- to Few-Shot
Sound
Helps computers better guess sounds using fewer examples.
Strategic Base Representation Learning via Feature Augmentations for Few-Shot Class Incremental Learning
CV and Pattern Recognition
Teaches computers new things without forgetting old ones.
Adapting Multimodal Foundation Models for Few-Shot Learning: A Comprehensive Study on Contrastive Captioners
CV and Pattern Recognition
Helps AI learn from very few pictures.