Towards Lossless Ultimate Vision Token Compression for VLMs
By: Dehua Zheng , Mouxiao Huang , Borui Jiang and more
Potential Business Impact:
Makes AI understand pictures much faster.
Visual language models encounter challenges in computational efficiency and latency, primarily due to the substantial redundancy in the token representations of high-resolution images and videos. Current attention/similarity-based compression algorithms suffer from either position bias or class imbalance, leading to significant accuracy degradation. They also fail to generalize to shallow LLM layers, which exhibit weaker cross-modal interactions. To address this, we extend token compression to the visual encoder through an effective iterative merging scheme that is orthogonal in spatial axes to accelerate the computation across the entire VLM. Furthermoer, we integrate a spectrum pruning unit into LLM through an attention/similarity-free low-pass filter, which gradually prunes redundant visual tokens and is fully compatible to modern FlashAttention. On this basis, we propose Lossless Ultimate Vision tokens Compression (LUVC) framework. LUVC systematically compresses visual tokens until complete elimination at the final layer of LLM, so that the high-dimensional visual features are gradually fused into the multimodal queries. The experiments show that LUVC achieves a 2 speedup inference in language model with negligible accuracy degradation, and the training-free characteristic enables immediate deployment across multiple VLMs.
Similar Papers
HybridToken-VLM: Hybrid Token Compression for Vision-Language Models
CV and Pattern Recognition
Lets computers understand pictures better, faster.
LLMC+: Benchmarking Vision-Language Model Compression with a Plug-and-play Toolkit
CV and Pattern Recognition
Makes AI understand pictures and words better, faster.
Fourier-VLM: Compressing Vision Tokens in the Frequency Domain for Large Vision-Language Models
CV and Pattern Recognition
Makes computers understand pictures much faster.