The official weight of NSA-1B (https://arxiv.org/abs/2502.11089) used in (https://arxiv.org/abs/2511.20102), serving as a baseline for SSA.
The sparse attention kernel is based on (https://github.com/fla-org/native-sparse-attention).
SSA: Sparse Sparse Attention by Aligning Full and Sparse Attention Outputs in Feature Space
Sparse attention reduces the quadratic complexity of full self-attention but faces two challenges: (1) an attention gap, where applying sparse attention to full-attention-trained models causes performance degradation due to train-inference distribution mismatch, and (2) a capability gap, where models trained purely with sparse attention lack complete gradient flow, preventing them from matching full-attention performance. We propose SSA (Sparse Sparse Attention), a training framework that integrates both sparse and full attention with bidirectional attention-output alignment. We prove that the approximation error scales linearly with the attention mass dropped under sparse attention, and show that SSA's alignment objective substantially reduces this quantity compared to baselines. Experiments demonstrate that SSA achieves state-of-the-art performance under both inference modes, adapts smoothly to varying sparsity budgets, and demonstrates superior long-context capabilities. The code is available at https://github.com/zhenyi4/ssa.
- Downloads last month
- 2