Score: 0

Distillation of a tractable model from the VQ-VAE

Published: September 1, 2025 | arXiv ID: 2509.01400v1

By: Armin Hadžić, Milan Papez, Tomáš Pevný

Potential Business Impact:

Makes AI understand and create better.

Business Areas:
Quantum Computing Science and Engineering

Deep generative models with discrete latent space, such as the Vector-Quantized Variational Autoencoder (VQ-VAE), offer excellent data generation capabilities, but, due to the large size of their latent space, their probabilistic inference is deemed intractable. We demonstrate that the VQ-VAE can be distilled into a tractable model by selecting a subset of latent variables with high probabilities. This simple strategy is particularly efficient, especially if the VQ-VAE underutilizes its latent space, which is, indeed, very often the case. We frame the distilled model as a probabilistic circuit, and show that it preserves expressiveness of the VQ-VAE while providing tractable probabilistic inference. Experiments illustrate competitive performance in density estimation and conditional generation tasks, challenging the view of the VQ-VAE as an inherently intractable model.

Page Count
12 pages

Category
Computer Science:
Machine Learning (CS)