HybridToken-VLM: Hybrid Token Compression for Vision-Language Models
By: Jusheng Zhang , Xiaoyang Guo , Kaitong Cai and more
Potential Business Impact:
Lets computers understand pictures better, faster.
Vision-language models (VLMs) have transformed multimodal reasoning, but feeding hundreds of visual patch tokens into LLMs incurs quadratic computational costs, straining memory and context windows. Traditional approaches face a trade-off: continuous compression dilutes high-level semantics such as object identities, while discrete quantization loses fine-grained details such as textures. We introduce HTC-VLM, a hybrid framework that disentangles semantics and appearance through dual channels, i.e., a continuous pathway for fine-grained details via ViT patches and a discrete pathway for symbolic anchors using MGVQ quantization projected to four tokens. These are fused into a 580-token hybrid sequence and compressed into a single voco token via a disentanglement attention mask and bottleneck, ensuring efficient and grounded representations. HTC-VLM achieves an average performance retention of 87.2 percent across seven benchmarks (GQA, VQAv2, MMBench, MME, POPE, SEED-Bench, ScienceQA-Image), outperforming the leading continuous baseline at 81.0 percent with a 580-to-1 compression ratio. Attention analyses show that the compressed token prioritizes the discrete anchor, validating its semantic guidance. Our work demonstrates that a minimalist hybrid design can resolve the efficiency-fidelity dilemma and advance scalable VLMs.
Similar Papers
LLMC+: Benchmarking Vision-Language Model Compression with a Plug-and-play Toolkit
CV and Pattern Recognition
Makes AI understand pictures and words better, faster.
Chain-of-Visual-Thought: Teaching VLMs to See and Think Better with Continuous Visual Tokens
CV and Pattern Recognition
Lets computers see and understand pictures better.
Fourier-VLM: Compressing Vision Tokens in the Frequency Domain for Large Vision-Language Models
CV and Pattern Recognition
Makes computers understand pictures much faster.