HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Wed, 17 Dec 2025 16:59:07 GMT
access-control-allow-origin: *
strict-transport-security: max-age=31556952
etag: W/"6942e15b-495d"
expires: Sun, 28 Dec 2025 13:30:58 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: 65A0:123DE:7A13CB:88E28F:69512EB8
accept-ranges: bytes
age: 0
date: Sun, 28 Dec 2025 13:20:58 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210054-BOM
x-cache: MISS
x-cache-hits: 0
x-timer: S1766928058.143139,VS0,VE228
vary: Accept-Encoding
x-fastly-request-id: 2e91ae7da7d8838839a39785ca10ca8205befd6f
content-length: 4285
Yang Chen
Skip to content
I am a Research Scientist at NVIDIA (ADLR Group). I received my Ph.D. from Georgia Tech in August 2024.
I currently work on scaling reinforcement learning for reasoning LLMs.
Selected Projects
Recent work
We scale Cascade RL to train general-purpose reasoning LLMs spanning RLHF, instruction following, math, code, and SWE.
Our 14B Thinking model achieves SOTA on LiveCodeBench—outperforming Gemini-2.5-Pro, o4-mini, Qwen3-235B, DeepSeek-R1-671B, and reaches silver-medal performance at IOI 2025.
We further scaled the SFT and studied the interplay with RL, and released the SOTA 7B model, AceReason-Nemotron-1.1-7B.
We scaled the previous work to both math and code domains and released the SOTA medium-sized model, AceReason-Nemotron-14B.
Our pilot study on scaling RL for competitive math reasoning LLMs.
We released the AceMath-RL-Nemotron-7B model.
We worked on developing math reasoning and reward models, and released AceMath-1.5B, 7B, and 72B,
along with AceMath-RM-7B and 72B, which surpassed GPT-4o and Qwen2.5-Math at the time.
Past Work
Research · 2022–2024
Multimodal LLM
Building AI that understands the visual world
Click to expand
Visual World Knowledge
[EMNLP'23 ,
ICCV'23 (Oral)]
Emergent Visual Privacy Concern
[EMNLP'24 (Oral)]
Universal Multimodal Retriever
[ECCV'24 (Oral)]
Research · 2019–2024
Multilingual LLM
Bridging representation across global languages
Click to expand
Model Selection
[EMNLP'21 ]
Cross-lingual
[ACL'23 Findings ,
ICLR'24 ,
ACL'25 ]
English–Arabic BERT
[EMNLP'20 ]
© Yang Chen
Simple, calm UI