DeepSeek's Next-Generation Technology Wins the ACL 2025 Best Paper Award, Enhancing Long-Text Processing Efficiency by 11 Times!
The paper by Dr. Liang Wenfeng's team at DeepSeek in collaboration with Peking University has won the ACL 2025 Best Paper Award. This research proposes the Native Sparse Attention (NSA) mechanism, which achieves algorithm-hardware co-optimization through three parallel attention branches: compression, selection, and sliding. It improves the processing speed of long texts by 11 times and outperforms traditional full-attention models. NSA supports a context length of up to 1 million tokens, achieving an 11.6 times decoding speed improvement on 64k text, and a 27B parameter model demonstrates excellent performance on multiple benchmark tests.