Feng Yao

Hi! My name is Feng (姚峰 in Chinese).
I am a second-year Ph.D. student in Computer Science at UCSD, advised by Prof. Jingbo Shang and Prof. Vish Krishnan. Previously, I received my master’s degree from Tsinghua University, advised by Prof. Zhiyuan Liu and Prof. Weixing Shen.
My research interest genearlly lies in the intersection of Natural Language Processing and Deep Learning. Recently, I have been focusing on training Mixture-of-Experts (MoE) and improving the efficiency of large-scale Reinforcement Learning.
Feel free to reach out if you want to collaborate with me. :)
News
Aug 12, 2025 | Invited by Kuaishou Klear Team to give a talk on MoE & RL topics. |
---|---|
Jul 14, 2025 | Invited by Cohere Labs to give a talk on DenseMixer. [Slides] |
Jul 02, 2025 | Invited by the Qwen Team to give a talk on DenseMixer. [Slides] |
Selected Publications [Full]
- Preprint
Experience
- Amazon Rufus Foundation Model Team | Jun 2025 – Sep 2025
Topic: Post-Training for LLM Agent
Hosts: Zheng Li, Xinyang Zhang, Changlong Yu, Shuowei Jin - Microsoft Research & GenAI | Jun 2024 – Mar 2025
Topic: MoE Pretraining / Reinforcement Learning
Hosts: Liyuan Liu, Yelong Shen, Shuohang Wang
Talks
- FlashRL & Your Efficient RL Framework Secretly Brings You Off-Policy RL Training
- @ Kuaishou Klear Team, August 19, 2025 [slides]
- DenseMixer: Improving MoE Post-Training with Precise Router Gradient.
- @ Kuaishou Klear Team, August 19, 2025 [slides]
- @ Cohere Lab, July 10, 2025 [slides] [recording]
- @ Alibaba Qwen Team, July 09, 2025 [slides]