LaMPE: Length-aware Multi-grained Positional Encoding for Adaptive Long-context Scaling Without Training
By: Sikui Zhang , Guangze Gao , Ziyun Gan and more
Potential Business Impact:
Lets computers understand much longer stories.
Large language models (LLMs) experience significant performance degradation when the input exceeds the pretraining context window, primarily due to the out-of-distribution (OOD) behavior of Rotary Position Embedding (RoPE). Recent studies mitigate this problem by remapping OOD positions into the in-distribution range with fixed mapping strategies, ignoring the dynamic relationship between input length and the model's effective context window. To this end, we propose Length-aware Multi-grained Positional Encoding (LaMPE), a training-free method that fully utilizes the model's effective context window for adaptive long-context scaling in LLMs. Motivated by the left-skewed frequency distribution of relative positions, LaMPE establishes a dynamic relationship between mapping length and input length through a parametric scaled sigmoid function to adaptively allocate positional capacity across varying input lengths. Meanwhile, LaMPE devises a novel multi-grained attention mechanism that strategically allocates positional resolution across different sequence regions to capture both fine-grained locality and long-range dependencies. Our method can be seamlessly applied to a wide range of RoPE-based LLMs without training. Extensive experiments on three representative LLMs across five mainstream long-context benchmarks demonstrate that LaMPE achieves significant performance improvements compared to existing length extrapolation methods. The code will be released at https://github.com/scar-on/LaMPE.
Similar Papers
LaMPE: Length-aware Multi-grained Position Encoding for Adaptive Long-context Scaling Without Training
Computation and Language
Lets AI understand much longer texts.
Layer-Specific Scaling of Positional Encodings for Superior Long-Context Modeling
Computation and Language
Helps AI remember important details in long texts.
Effective Length Extrapolation via Dimension-Wise Positional Embeddings Manipulation
Computation and Language
Lets computers remember much longer stories.