LASP-2: Rethinking Sequence Parallelism for Linear Attention and Its HybridFeb 11, 2025·Weigao Sun,Disen Lan,Yiran Zhong,Xiaoye Qu,Yu Cheng· 0 min read PDF Cite CodeLast updated on Feb 11, 2025 AuthorsWeigao SunYoung Scientist ← MoM: Linear Sequence Modeling with Mixture-of-Memories Feb 19, 2025Minimax-01: Scaling Foundation Models with Lightning Attention Jan 14, 2025 →