A Comprehensive Study on Visual Token Redundancy for Discrete Diffusion-based Multimodal Large Language Models
By: Duo Li , Zuhao Yang , Xiaoqin Zhang and more
Potential Business Impact:
Makes AI understand pictures and words much faster.
Discrete diffusion-based multimodal large language models (dMLLMs) have emerged as a promising alternative to autoregressive MLLMs thanks to their advantages in parallel decoding and bidirectional context modeling, but most existing dMLLMs incur significant computational overhead during inference due to the full-sequence attention computation in each denoising step. Pioneer studies attempt to resolve this issue from a modality-agnostic perspective via key-value cache optimization or efficient sampling but most of them overlook modality-specific visual token redundancy. In this work, we conduct a comprehensive study on how visual token redundancy evolves with different dMLLM architectures and tasks and how visual token pruning affects dMLLM responses and efficiency. Specifically, our study reveals that visual redundancy emerges only in from-scratch dMLLMs while handling long-answer tasks. In addition, we validate that visual token pruning introduces non-negligible information loss in dMLLMs and only from-scratch dMLLMs can recover the lost information progressively during late denoising steps. Furthermore, our study shows that layer-skipping is promising for accelerating AR-to-diffusion dMLLMs, whereas progressive or late-step pruning is more effective for from-scratch dMLLMs. Overall, this work offers a new perspective on efficiency optimization for dMLLMs, greatly advancing their applicability across various multimodal understanding tasks.
Similar Papers
Towards Adaptive Visual Token Pruning for Large Multimodal Models
CV and Pattern Recognition
Makes AI understand pictures faster and cheaper.
Discrete Diffusion in Large Language and Multimodal Models: A Survey
Machine Learning (CS)
Makes AI talk and create much faster.
WeMMU: Enhanced Bridging of Vision-Language Models and Diffusion Models via Noisy Query Tokens
CV and Pattern Recognition
Teaches AI to learn many new things without forgetting.