Autoregressive Speech Enhancement via Acoustic Tokens
By: Luca Della Libera, Cem Subakan, Mirco Ravanelli
Potential Business Impact:
Makes voices clearer by using sound pieces.
In speech processing pipelines, improving the quality and intelligibility of real-world recordings is crucial. While supervised regression is the primary method for speech enhancement, audio tokenization is emerging as a promising alternative for a smooth integration with other modalities. However, research on speech enhancement using discrete representations is still limited. Previous work has mainly focused on semantic tokens, which tend to discard key acoustic details such as speaker identity. Additionally, these studies typically employ non-autoregressive models, assuming conditional independence of outputs and overlooking the potential improvements offered by autoregressive modeling. To address these gaps we: 1) conduct a comprehensive study of the performance of acoustic tokens for speech enhancement, including the effect of bitrate and noise strength; 2) introduce a novel transducer-based autoregressive architecture specifically designed for this task. Experiments on VoiceBank and Libri1Mix datasets show that acoustic tokens outperform semantic tokens in terms of preserving speaker identity, and that our autoregressive approach can further improve performance. Nevertheless, we observe that discrete representations still fall short compared to continuous ones, highlighting the need for further research in this area.
Similar Papers
Modeling strategies for speech enhancement in the latent space of a neural audio codec
Sound
Makes noisy speech clear by learning its hidden sounds.
Speech Tokenizer is Key to Consistent Representation
Machine Learning (CS)
Makes computers understand talking better, even feelings.
Parallel GPT: Harmonizing the Independence and Interdependence of Acoustic and Semantic Information for Zero-Shot Text-to-Speech
Audio and Speech Processing
Makes computer voices sound more human and real.