Research Scientist @ NVIDIA
I am an AI Research Scientist @ NVIDIA (ADLR Group) building Large Language Models. I received my Ph.D. from Georgia Tech in August 2024.
My current research interests lie in AI reasoning and scaling reinforcement learning for LLM reasoning.
Announcements
AceReason: Math and Code Reasoning through RL
TLDR: A math and code reasoning model trained entirely through RL. We found math-RL improves code reasoning and use it as a warm-up for code-RL to improve training stability and efficiency.
May 22, 2025
Announcements
AceMath-RL: Training Math Reasoning Model with RL
TLDR: A math reasoning model trained entirely through RL, starting from the Deepseek-R1-Distilled-Qwen-7B.
We share our training recipe, training logs, and data curation details in the blog.
April 22, 2025
Announcements
AceMath: Advancing Frontier Math Reasoning with Post-Training and Reward Modeling
Hugginface (Models & Datasets)
Technical Report
Project Page
Dec 19th, 2024
Multimodal LLM
Building AI that understand visual world
Multilingual LLM
Bridging representation across global languages