Songlin Yang

profile.jpg

Songlin (松琳) is a Member of Technical Staff at Thinking Machines Lab, working on language model architectures. She earned her PhD from MIT, where she was advised by Prof. Yoon Kim.


Flash Linear Attention efficient attention implementations in Triton
FLA Discord community for Flash Linear Attention
ASAP Seminar Advances in Sequence Modeling from Algorithmic Perspectives

latest posts

selected publications

  1. ICML
    Gated Linear Attention Transformers with Hardware-Efficient Training
    Songlin Yang*, Bailin Wang*Yikang ShenRameswar Panda, and Yoon Kim
    In , 2024
  2. ICLR
    Gated Delta Networks: Improving Mamba2 with Delta Rule
    Songlin Yang, Jan Kautz, and Ali Hatamizadeh
    2025
  3. NeurIPS
    Parallelizing Linear Transformers with the Delta Rule over Sequence Length
    Songlin Yang, Bailin WangYu ZhangYikang Shen, and Yoon Kim
    In , 2024
  4. NeurIPS
    PaTH Attention: Position Encoding via Accumulating Householder Transformations
    2025