Adaptive Token Merging for Efficient Transformer Semantic Communication at the Edge
By: Omar Erak , Omar Alhussein , Hatem Abou-Zeid and more
Potential Business Impact:
Makes smart computer programs run faster, cheaper.
Large-scale transformers are central to modern semantic communication, yet their high computational and communication costs hinder deployment on resource-constrained edge devices. This paper introduces a training-free framework for adaptive token merging, a novel mechanism that compresses transformer representations at runtime by selectively merging semantically redundant tokens under per-layer similarity thresholds. Unlike prior fixed-ratio reduction, our approach couples merging directly to input redundancy, enabling data-dependent adaptation that balances efficiency and task relevance without retraining. We cast the discovery of merging strategies as a multi-objective optimization problem and leverage Bayesian optimization to obtain Pareto-optimal trade-offs between accuracy, inference cost, and communication cost. On ImageNet classification, we match the accuracy of the unmodified transformer with 30\% fewer floating-point operations per second and under 20\% of the original communication cost, while for visual question answering our method achieves performance competitive with the full LLaVA model at less than one-third of the compute and one-tenth of the bandwidth. Finally, we show that our adaptive merging is robust across varying channel conditions and provides inherent privacy benefits, substantially degrading the efficacy of model inversion attacks. Our framework provides a practical and versatile solution for deploying powerful transformer models in resource-limited edge intelligence scenarios.
Similar Papers
Adaptive Pareto-Optimal Token Merging for Edge Transformer Models in Semantic Communication
Machine Learning (CS)
Lets AI understand pictures faster on phones.
AdaTok: Adaptive Token Compression with Object-Aware Representations for Efficient Multimodal LLMs
CV and Pattern Recognition
Makes AI understand pictures using fewer computer steps.
Token Transforming: A Unified and Training-Free Token Compression Framework for Vision Transformer Acceleration
CV and Pattern Recognition
Makes AI see faster with less work.