Score: 1

QINCODEC: Neural Audio Compression with Implicit Neural Codebooks

Published: March 19, 2025 | arXiv ID: 2503.19597v1

By: Zineb Lahrichi , Gaëtan Hadjeres , Gael Richard and more

Potential Business Impact:

Makes audio sound better with simpler training.

Business Areas:
Quantum Computing Science and Engineering

Neural audio codecs, neural networks which compress a waveform into discrete tokens, play a crucial role in the recent development of audio generative models. State-of-the-art codecs rely on the end-to-end training of an autoencoder and a quantization bottleneck. However, this approach restricts the choice of the quantization methods as it requires to define how gradients propagate through the quantizer and how to update the quantization parameters online. In this work, we revisit the common practice of joint training and propose to quantize the latent representations of a pre-trained autoencoder offline, followed by an optional finetuning of the decoder to mitigate degradation from quantization. This strategy allows to consider any off-the-shelf quantizer, especially state-of-the-art trainable quantizers with implicit neural codebooks such as QINCO2. We demonstrate that with the latter, our proposed codec termed QINCODEC, is competitive with baseline codecs while being notably simpler to train. Finally, our approach provides a general framework that amortizes the cost of autoencoder pretraining, and enables more flexible codec design.

Page Count
5 pages

Category
Computer Science:
Sound