Songlin Yang

sonta.jpg

Songlin is currently a first-year PhD student at MIT CSAIL, advised by Prof. Yoon Kim.

Previously, she obtained her bachelar’s degree from SUSTech in 2020 and her master’s degree from ShanghaiTech in 2023, where she was advised by Prof. Kewei Tu.

Her research is centered on the intersection of machine learning system and large language model, with a specific focus on the hardware-aware algorithm design for efficient sequence modeling.

news

Apr 12, 2024 Introducing HGRN2 :sparkles: :sparkles:, a minimalist linear attention model with strong performance. Code is available at here.
Jan 1, 2024 Introducing our open-source project flash-linear-attention :rocket: :rocket: :rocket:. Join Discord if you are interested in linear attention/RNN!
Dec 14, 2023 Presenting our linear RNN paper (HGRN) at NeurIPS :sparkles: :sparkles:. Code is available at here.
Dec 12, 2023 Announcing Gated Linear Attention Transformers (GLA)! :sparkles: :smile: Code is available at here.

selected publications

  1. arXiv
    HGRN2: Gated Linear RNNs with State Expansion
    Zhen Qin, Songlin Yang, Weixuan Sun, Xuyang Shen, Dong Li, Weigao Sun, and Yiran Zhong
    In , 2024
  2. arXiv
    Gated Linear Attention Transformers with Hardware-Efficient Training
    Songlin Yang*, Bailin Wang*Yikang ShenRameswar Panda, and Yoon Kim
    In , 2023
  3. NeurIPS spotlight
    Hierarchically Gated Recurrent Neural Network for Sequence Modeling
    Zhen Qin*, Songlin Yang*, and Yiran Zhong
    In Thirty-seventh Conference on Neural Information Processing Systems, 2023