Score: 1

Accelerate Speculative Decoding with Sparse Computation in Verification

Published: December 26, 2025 | arXiv ID: 2512.21911v1

By: Jikai Wang , Jianchao Tan , Yuxuan Hu and more

BigTech Affiliations: Meituan

Potential Business Impact:

Makes AI write faster without losing quality.

Business Areas:
Semantic Search Internet Services

Speculative decoding accelerates autoregressive language model inference by verifying multiple draft tokens in parallel. However, the verification stage often becomes the dominant computational bottleneck, especially for long-context inputs and mixture-of-experts (MoE) models. Existing sparsification methods are designed primarily for standard token-by-token autoregressive decoding to remove substantial computational redundancy in LLMs. This work systematically adopts different sparse methods on the verification stage of the speculative decoding and identifies structured redundancy across multiple dimensions. Based on these observations, we propose a sparse verification framework that jointly sparsifies attention, FFN, and MoE components during the verification stage to reduce the dominant computation cost. The framework further incorporates an inter-draft token and inter-layer retrieval reuse strategy to further reduce redundant computation without introducing additional training. Extensive experiments across summarization, question answering, and mathematical reasoning datasets demonstrate that the proposed methods achieve favorable efficiency-accuracy trade-offs, while maintaining stable acceptance length.

Country of Origin
🇨🇳 China

Page Count
11 pages

Category
Computer Science:
Computation and Language