Songlin Yang

profile.jpg

Songlin (松琳) is a final-year PhD student at MIT CSAIL, advised by Prof. Yoon Kim. She is also a Member of Technical Staff at Thinking Machines Lab.

Her research focuses on language model architectures, in particular attention mechanisms. To learn more about her work, see talks.


Flash Linear Attention efficient attention implementations in Triton
FLA Discord community for Flash Linear Attention
ASAP Seminar Advances in Sequence Modeling from Algorithmic Perspectives

latest posts

selected publications

  1. ICML
    Gated Linear Attention Transformers with Hardware-Efficient Training
    Songlin Yang*, Bailin Wang*Yikang ShenRameswar Panda, and Yoon Kim
    In , 2024
  2. ICLR
    Gated Delta Networks: Improving Mamba2 with Delta Rule
    Songlin Yang, Jan Kautz, and Ali Hatamizadeh
    2025
  3. NeurIPS
    Parallelizing Linear Transformers with the Delta Rule over Sequence Length
    Songlin Yang, Bailin WangYu ZhangYikang Shen, and Yoon Kim
    In , 2024
  4. NeurIPS
    PaTH Attention: Position Encoding via Accumulating Householder Transformations
    2025