HTTP/2 200
server: GitHub.com
content-type: text/html; charset=utf-8
last-modified: Mon, 01 Dec 2025 21:14:41 GMT
access-control-allow-origin: *
etag: W/"692e0541-5038"
expires: Sun, 28 Dec 2025 19:13:20 GMT
cache-control: max-age=600
content-encoding: gzip
x-proxy-cache: MISS
x-github-request-id: 74D9:3A7A40:7E5E67:8DC832:69517EF8
accept-ranges: bytes
age: 0
date: Sun, 28 Dec 2025 19:16:59 GMT
via: 1.1 varnish
x-served-by: cache-bom-vanm7210025-BOM
x-cache: HIT
x-cache-hits: 0
x-timer: S1766949419.876556,VS0,VE198
vary: Accept-Encoding
x-fastly-request-id: 69d2ced6b3146c9d6f9bce5058a1440e56e1570f
content-length: 5902
Tri Dao
tri [at] tridao (dot) me
Assistant Professor of Computer Science at Princeton University.
Co-founder & Chief Scientist of Together AI.
CV (updated 12/2025)
Previously: PhD, Department of Computer Science, Stanford University
Research Interests
Machine learning and systems, with a focus on efficient training and inference:
- Hardware-aware algorithms.
- Sequence models with long-range memory.
Current PhD Students
Selected Honors and Awards
- Schmidt Sciences AI2050 Fellowship, 2025.
- Google ML and Systems Junior Faculty Awards, 2025.
- Google Research Scholar, 2025.
- Conference on Machine Learning and Systems (MLSys), Outstanding Paper Honorable Mention, 2025.
- Conference on Language Modeling (COLM), Outstanding Paper, 2024.
- International Conference on Machine Learning (ICML), Outstanding Paper runner-up, 2022.
-
Marconi: Prefix Caching for the Era of Hybrid LLMs
Rui Pan, Zhuang Wang, Zhen Jia, Can Karakus, Luca Zancato, Tri Dao, Ravi Netravali, and Yida Wang
In Machine Learning and Systems (MLSys) , 2025
Outstanding Paper Honorable Mention
-
FlashAttention-3: Fast and Accurate Attention with Asynchrony and Low-precision
Jay Shah*, Ganesh Bikshandi*, Ying Zhang, Vijay Thakkar, Pradeep Ramani, and Tri Dao
In Advances in Neural Information Processing Systems (NeurIPS) , 2024
-
Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
Tri Dao*, and Albert Gu*
In International Conference on Machine Learning (ICML) , 2024
-
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Albert Gu*, and Tri Dao*
Conference on Language Modeling (COLM), 2023
-
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
Tri Dao, Daniel Y. Fu, Stefano Ermon, Atri Rudra, and Christopher Ré
In Advances in Neural Information Processing Systems , 2022
Best Paper award at the ICML Hardware Aware Efficient Training Workshop 2022, Inaugural Stanford Open Source Software Prize 2024
-
Monarch: Expressive Structured Matrices for Efficient and Accurate Training
Tri Dao, Beidi Chen, Nimit Sohoni, Arjun Desai, Michael Poli, Jessica Grogan, Alexander Liu, Aniruddh Rao, Atri Rudra, and Christopher Ré
In International Conference on Machine Learning (ICML) , 2022
Outstanding Paper runner-up