Feng Yao

Hi! My name is Feng (姚峰 in Chinese).
I am a second-year Ph.D. student in Computer Science at UCSD, advised by Prof. Jingbo Shang and Prof. Vish Krishnan. Previously, I received my master’s degree from Tsinghua University, advised by Prof. Zhiyuan Liu and Prof. Weixing Shen.
My research interest genearlly lies in the intersection of Natural Language Processing and Deep Learning. Recently, I have been focusing on training Mixture-of-Experts (MoE) and improving the efficiency of large-scale Reinforcement Learning.
Feel free to reach out if you want to collaborate with me. :)
News
Aug 25, 2025 | Invited by MiniMax and Sea AI Lab to give a talk on DenseMixer, TIS & FlashRL. |
---|---|
Aug 24, 2025 | Invited by TsinghuaNLP and ModelBest to give a talk on TIS & FlashRL. [Slides] |
Aug 17, 2025 | |
Aug 12, 2025 | Invited by Kuaishou Klear Team to give a talk on DenseMixer, TIS & FlashRL. |
Jul 14, 2025 | Invited by Cohere Labs to give a talk on DenseMixer. [Slides] |
Jul 02, 2025 | Invited by the Qwen Team to give a talk on DenseMixer. [Slides] |
Jun 30, 2025 | Released DenseMixer for MoE post-training! Check out the blog, code and X. |
Selected Publications [Full]
- Preprint
Experience
- Amazon Rufus Foundation Model Team | Jun 2025 – Sep 2025
Topic: Post-Training for LLM Agent
Hosts: Zheng Li, Changlong Yu, Shuowei Jin, Lihong Li - Microsoft Research & GenAI | Jun 2024 – Mar 2025
Topic: MoE Training / Reinforcement Learning
Hosts: Liyuan Liu, , Shuohang Wang, Yelong Shen, , Jianfeng Gao
Invited Talks
- On the Rollout-Training Mismatch in Modern RL Systems
- DenseMixer: Improving MoE Post-Training with Precise Router Gradient.
- @ MiniMax, September 01, 2025 [slides]
- @ Kuaishou Klear Team, August 19, 2025 [slides]
- @ Cohere Lab, July 10, 2025 [slides] [recording]
- @ Alibaba Qwen Team, July 09, 2025 [slides]