Pretrained Conformers for Audio Fingerprinting and Retrieval
By: Kemal Altwlkany, Elmedin Selmanovic, Sead Delalic
Potential Business Impact:
Finds sounds in noisy recordings quickly.
Conformers have shown great results in speech processing due to their ability to capture both local and global interactions. In this work, we utilize a self-supervised contrastive learning framework to train conformer-based encoders that are capable of generating unique embeddings for small segments of audio, generalizing well to previously unseen data. We achieve state-of-the-art results for audio retrieval tasks while using only 3 seconds of audio to generate embeddings. Our models are almost completely immune to temporal misalignments and achieve state-of-the-art results in cases of other audio distortions such as noise, reverb or extreme temporal stretching. Code and models are made publicly available and the results are easy to reproduce as we train and test using popular and freely available datasets of different sizes.
Similar Papers
Pretrained Conformers for Audio Fingerprinting and Retrieval
Sound
Finds sounds in noisy recordings quickly.
Unveiling Audio Deepfake Origins: A Deep Metric learning And Conformer Network Approach With Ensemble Fusion
Sound
Finds who made fake voices.
Conformer-based Ultrasound-to-Speech Conversion
Sound
Turns silent mouth movements into speech.