Score: 0

Group-Wise Optimization for Self-Extensible Codebooks in Vector Quantized Models

Published: October 15, 2025 | arXiv ID: 2510.13331v1

By: Hong-Kai Zheng, Piji Li

Potential Business Impact:

Improves AI image making by better organizing its drawing tools.

Business Areas:
Image Recognition Data and Analytics, Software

Vector Quantized Variational Autoencoders (VQ-VAEs) leverage self-supervised learning through reconstruction tasks to represent continuous vectors using the closest vectors in a codebook. However, issues such as codebook collapse persist in the VQ model. To address these issues, existing approaches employ implicit static codebooks or jointly optimize the entire codebook, but these methods constrain the codebook's learning capability, leading to reduced reconstruction quality. In this paper, we propose Group-VQ, which performs group-wise optimization on the codebook. Each group is optimized independently, with joint optimization performed within groups. This approach improves the trade-off between codebook utilization and reconstruction performance. Additionally, we introduce a training-free codebook resampling method, allowing post-training adjustment of the codebook size. In image reconstruction experiments under various settings, Group-VQ demonstrates improved performance on reconstruction metrics. And the post-training codebook sampling method achieves the desired flexibility in adjusting the codebook size.

Country of Origin
🇨🇳 China

Page Count
18 pages

Category
Computer Science:
CV and Pattern Recognition