Index-Preserving Lightweight Token Pruning for Efficient Document Understanding in Vision-Language Models
By: Jaemin Son, Sujin Choi, Inyong Yun
Potential Business Impact:
Makes AI understand papers faster and cheaper.
Recent progress in vision-language models (VLMs) has led to impressive results in document understanding tasks, but their high computational demands remain a challenge. To mitigate the compute burdens, we propose a lightweight token pruning framework that filters out non-informative background regions from document images prior to VLM processing. A binary patch-level classifier removes non-text areas, and a max-pooling refinement step recovers fragmented text regions to enhance spatial coherence. Experiments on real-world document datasets demonstrate that our approach substantially lowers computational costs, while maintaining comparable accuracy.
Similar Papers
PoRe: Position-Reweighted Visual Token Pruning for Vision Language Models
CV and Pattern Recognition
Helps AI focus on important parts of pictures.
Training-Free Pyramid Token Pruning for Efficient Large Vision-Language Models via Region, Token, and Instruction-Guided Importance
CV and Pattern Recognition
Focuses on important image parts for faster AI.
VLM-Pruner: Buffering for Spatial Sparsity in an Efficient VLM Centrifugal Token Pruning Paradigm
CV and Pattern Recognition
Makes AI understand pictures faster on phones.