I am a final-year PhD student from the Department of Computer Science and Technology, Tsinghua University. I am fortunately supervised by Professor Maosong Sun and Zhiyuan Liu. In the summer of 2019, I visited MILA and conducted research under Professor Jian Tang.

My current passion revolves around building SCALABLE solutions to ASI (AGI is too vague to be a target, and its first version will be achieved soon.), which means the solutions will bring improvement simply with more resources on computation and data. This include:

  1. Scaling Pretrain. Ensuring the growth of language models’ ability is measureable and predictable, and potentially instructing us to find better scaling law with better paradigm.

  2. Scaling RL. This includes developing scaling RL training compute, inference compute, data source, RL horizon (O1 paradigm).

  3. Scaling World Model Unifing modality and training objectives, then scaling with pretrain and RL.

The above objects seem very broad. I will focus on one point at a time. Currently, I am on ``Better Architecture for RL Scaling’'.

In fact, I believe that we will not have achieved ASI until the model is capable of conducting scientific research independently. There is still a long way to go. Keep fighting!

Selected Publications

For other papers, please refer to my google scholar

Selected Projects

Readme Card

Readme Card

Readme Card

Readme Card

Readme Card

Readme Card

Awards

Master&PhD

  • National Natural Science Foundation of China (NSFC) Doctoral Project Leader (~1/100 among all subjects and all institutes in China)

  • Siebel Scholar of Class 2023

    1/83 across the world, Price $30,000

  • National Scholarship 2021.

    One of the highest award in Tsinghua University.

Bachelor

  • Academic Excellence Award in 2016-2017 , 2017-2018, 2018-2019.
  • Good reading scholarship in 2018-2019 year.
  • Zhuzhou Scholarship in 2017-2018 year

Earlier

  • Silver Medal in the 32nd National Middle School Physics Competition