EffiVLM-BENCH: A Comprehensive Benchmark for Evaluating Training-Free Acceleration in Large Vision-Language Models
By: Zekun Wang , Minghua Ma , Zexin Wang and more
Potential Business Impact:
Makes big AI models run much faster and cheaper.
Large Vision-Language Models (LVLMs) have achieved remarkable success, yet their significant computational demands hinder practical deployment. While efforts to improve LVLM efficiency are growing, existing methods lack comprehensive evaluation across diverse backbones, benchmarks, and metrics. In this work, we systematically evaluate mainstream acceleration techniques for LVLMs, categorized into token and parameter compression. We introduce EffiVLM-Bench, a unified framework for assessing not only absolute performance but also generalization and loyalty, while exploring Pareto-optimal trade-offs. Our extensive experiments and in-depth analyses offer insights into optimal strategies for accelerating LVLMs. We open-source code and recipes for EffiVLM-Bench to foster future research.
Similar Papers
LVLM-Compress-Bench: Benchmarking the Broader Impact of Large Vision-Language Model Compression
CV and Pattern Recognition
Makes AI that sees and talks smaller, faster.
Resampling Benchmark for Efficient Comprehensive Evaluation of Large Vision-Language Models
CV and Pattern Recognition
Tests smart AI faster and more fairly.
Benchmarking Energy Efficiency of Large Language Models Using vLLM
Software Engineering
Helps make AI use less electricity.