Optimizing Native Sparse Attention with Latent Attention and Local Global Alternating Strategies
By: Yuxuan Hu , Jianchao Tan , Jiaqi Zhang and more
Potential Business Impact:
Makes computers understand long stories better.
In this work, we conduct a systematic analysis of Native Sparse Attention (NSA) and propose targeted improvements that enhance long-context modeling. A key insight is that alternating between local (sliding-window) and global (compression, selective) attention across layers, rather than using fixed patterns, enables more effective propagation of long-range dependencies and substantially boosts performance on long-sequence tasks. Meanwhile, we further refine NSA's branches with Latent Attention that the sliding-window branch is enhanced with Multi-head Latent Attention (MLA) while compression and selective branches adopt Group-head Latent Attention (GLA). These changes reduce KV-cache memory by 50\% versus NSA while improving the model's common-sense reasoning and long-text understanding capabilities. Experiments on models from 340M to 1.3B parameters (trained on 15B and 100B tokens) show our method matches or exceeds full attention and native sparse attention in both common-sense reasoning and long-context understanding tasks.
Similar Papers
SSA: Sparse Sparse Attention by Aligning Full and Sparse Attention Outputs in Feature Space
Computation and Language
Makes AI understand long stories better, faster.
Flash Sparse Attention: An Alternative Efficient Implementation of Native Sparse Attention Kernel
Distributed, Parallel, and Cluster Computing
Makes AI understand more words faster.
VideoNSA: Native Sparse Attention Scales Video Understanding
CV and Pattern Recognition
Lets computers watch and understand long videos better.