Accelerate Speculative Decoding with Sparse Computation in Verification
By: Jikai Wang , Jianchao Tan , Yuxuan Hu and more
Potential Business Impact:
Makes AI write faster without losing quality.
Speculative decoding accelerates autoregressive language model inference by verifying multiple draft tokens in parallel. However, the verification stage often becomes the dominant computational bottleneck, especially for long-context inputs and mixture-of-experts (MoE) models. Existing sparsification methods are designed primarily for standard token-by-token autoregressive decoding to remove substantial computational redundancy in LLMs. This work systematically adopts different sparse methods on the verification stage of the speculative decoding and identifies structured redundancy across multiple dimensions. Based on these observations, we propose a sparse verification framework that jointly sparsifies attention, FFN, and MoE components during the verification stage to reduce the dominant computation cost. The framework further incorporates an inter-draft token and inter-layer retrieval reuse strategy to further reduce redundant computation without introducing additional training. Extensive experiments across summarization, question answering, and mathematical reasoning datasets demonstrate that the proposed methods achieve favorable efficiency-accuracy trade-offs, while maintaining stable acceptance length.
Similar Papers
Accelerating Large-Scale Reasoning Model Inference with Sparse Self-Speculative Decoding
Machine Learning (CS)
Makes AI answer questions much faster.
Confidence-Modulated Speculative Decoding for Large Language Models
Computation and Language
Makes AI write faster and smarter.
Scaling LLM Speculative Decoding: Non-Autoregressive Forecasting in Large-Batch Scenarios
Computation and Language
Makes AI write faster without wasting power.