Distillation of a tractable model from the VQ-VAE
By: Armin Hadžić, Milan Papez, Tomáš Pevný
Potential Business Impact:
Makes AI understand and create better.
Deep generative models with discrete latent space, such as the Vector-Quantized Variational Autoencoder (VQ-VAE), offer excellent data generation capabilities, but, due to the large size of their latent space, their probabilistic inference is deemed intractable. We demonstrate that the VQ-VAE can be distilled into a tractable model by selecting a subset of latent variables with high probabilities. This simple strategy is particularly efficient, especially if the VQ-VAE underutilizes its latent space, which is, indeed, very often the case. We frame the distilled model as a probabilistic circuit, and show that it preserves expressiveness of the VQ-VAE while providing tractable probabilistic inference. Experiments illustrate competitive performance in density estimation and conditional generation tasks, challenging the view of the VQ-VAE as an inherently intractable model.
Similar Papers
An Introduction to Discrete Variational Autoencoders
Machine Learning (CS)
Teaches computers to understand words by grouping them.
Vector Quantization using Gaussian Variational Autoencoder
Machine Learning (CS)
Makes images easier for computers to understand.
Interpretable representation learning of quantum data enabled by probabilistic variational autoencoders
Quantum Physics
Finds hidden patterns in quantum data automatically.