Avatar

Liyuan Liu

Principal Researcher @ MSR

Hi there!

Welcome to Liyuan Lucas Liu (刘力源)‘s webpage! I am a Principal Researcher at Microsoft Research. My Ph.D. advisor is Prof. Jiawei Han, and my undergraduate advisor is Prof. Linli Xu. My research is about to understand the underlying mechanism of pretraining heuristics.

If you are going to visit Redmond, please let me buy you an ice cream. Molly Moon and Salt and Straw are pretty good .

Experience

  • Principal Researcher

    MSR, 2025 → Present

  • Senior Researcher

    MSR, 2022 → 2025

Education

  • Ph.D. in Computer Science

    University of Illinois at Urbana-Champaign

  • B.Eng. in Computer Science

    University of Science and Technology of China

Things I do

... and want to do

I mainly work on understanding the mechanisms behind common heuristics in machine learning (aka tricks). Some interesting ones:


Fun Facts about Me

Love skiing & met my amazing wife during a ski trip; DJI fans; play Sheng Ji (双扣), Ark Nova, and Mafia (狼人杀) with families & friends.

Selected Publications

List of all publications >>

(2023). Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs. Proceedings of the Twelfth International Conference on Learning Representations (ICLR 2024). Outstanding Paper Honorable Mention.

PDF

(2023). Bridging Discrete and Backpropagation: Straight-Through and Beyond. Proceedings of the Proceeding of Thirty-seventh Annual Conference on Neural Information Processing Systems (NeurIPS 2023). Selected as Oral.

PDF

(2022). Label Noise in Adversarial Training: A Novel Perspective to Study Robust Overfitting. Proceedings of the Thirty-sixth Annual Conference on Neural Information Processing Systems (NeurIPS 2022). Selected as Oral.

PDF

(2020). Understanding the Difficulty of Training Transformers. the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP 2020). Selected as Oral.

PDF Code Slide

(2020). On the Variance of the Adaptive Learning Rate and Beyond. the Eighth International Conference on Learning Representations (ICLR 2020).

PDF Code

(2018). Efficient Contextualized Representation: Language Model Pruning for Sequence Labeling. the 2018 Conference on Empirical Methods in Natural Language Processing (EMNLP 2018). Selected as Oral.

PDF Code Blog Doc

(2018). Empower Sequence Labeling with Task-Aware Neural Language Model. Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence (AAAI 2018).

PDF Code (new) Code (old) Blog

Highlighted Honors

List of all honors »

Outstanding Paper Honorable Mention

Shortlisted as 1 of 16 papers among 4922 submissions.

Winner of the Topcoder Arabic NER Challenge

Ranked 1st among 137 registrants and 220 submissions.

Guo Moruo Scholarship

Highest honor for USTC undergraduate students.

Google Excellent Scholarship

Only 58 graduate and undergraduate students shortlisted nationwide.

Contact

… stay in touch!