Group-Wise Optimization for Self-Extensible Codebooks in Vector Quantized Models
By: Hong-Kai Zheng, Piji Li
Potential Business Impact:
Improves AI image making by better organizing its drawing tools.
Vector Quantized Variational Autoencoders (VQ-VAEs) leverage self-supervised learning through reconstruction tasks to represent continuous vectors using the closest vectors in a codebook. However, issues such as codebook collapse persist in the VQ model. To address these issues, existing approaches employ implicit static codebooks or jointly optimize the entire codebook, but these methods constrain the codebook's learning capability, leading to reduced reconstruction quality. In this paper, we propose Group-VQ, which performs group-wise optimization on the codebook. Each group is optimized independently, with joint optimization performed within groups. This approach improves the trade-off between codebook utilization and reconstruction performance. Additionally, we introduce a training-free codebook resampling method, allowing post-training adjustment of the codebook size. In image reconstruction experiments under various settings, Group-VQ demonstrates improved performance on reconstruction metrics. And the post-training codebook sampling method achieves the desired flexibility in adjusting the codebook size.
Similar Papers
Group-Wise Optimization for Self-Extensible Codebooks in Vector Quantized Models
CV and Pattern Recognition
Improves AI's picture making and editing skills.
Scalable Training for Vector-Quantized Networks with 100% Codebook Utilization
CV and Pattern Recognition
Makes AI draw better pictures by fixing image parts.
Hierarchical Vector Quantized Graph Autoencoder with Annealing-Based Code Selection
Machine Learning (CS)
Teaches computers to understand connections in data.