I am an assistant professor in the Decision, Risk, and Operations division at Columbia Business School and a member of the Data Science Institute. I work on building trustworthy AI systems that are capable of
solving real-world decision-making problems. I take a data-centric view of AI systems, and am a strong believer in algorithmic ideas
simultaneously grounded in empirical foundations and principled thinking.
As an interdisciplinary researcher, I connect and extend tools from machine learning, operations research, and statistics.
Read this overview of my research to learn more about my impact-driven agenda.
Before joining Columbia, I received my Ph.D. from Stanford University and spent a year at Meta’s Adaptive Experimentation team as a research scientist. Outside of academia, I serve as a LinkedIn Scholar at LinkedIn’s Core AI team.
Here’s a more formal bio in the third person. I go by Hong; alternatively, here’s a link the correct pronunciation of my first name.
I’m looking for motivated undergraduate and master’s students to work on ML research. Fill this form out if you’re interested.
Dec 05, 2024
AI systems are omni-present, yet extrapolate unreliably. Improving AI safety and capabilities
hinges on comprehension of uncertainty and actively making decisions to resolve it.
Instead of cumbersome probabilistic models, my team leverages a predictive view of uncertainty
to build a scalable framework based on autoregressive models. Watch this recent Simons talk to learn more.
@article{MittalZhDoNa25,title={Data-driven Stochastic Modeling using Autoregressive Sequence Models},author={Mittal, Daksh and Zheng, Shunri and Dong, Jing and Namkoong, Hongseok},journal={arxiv:2509.05839 [cs.LG]},year={2025},url={https://arxiv.org/abs/2509.05839}}
Selected for oral presentations at the Econometric Society
Interdisciplinary Frontiers: Economics and AI+ML
conference and Conference on Digital Experimentation
Adaptivity
can significantly improve efficiency of experimentation, but it is challenging to implement even at large
online platforms with mature experimentation systems.
As a result, many real-world
experiments are deliberately implemented with large batches and a handful of
opportunities to update the sampling allocation as a way to reduce operational
costs of experimentation.
In this work, we focus on adaptive experiments with limited adaptivity (short horizons T < 10). Bandit algorithms focusing on long-horizon settings are tailored to provide regret guarantees for each specific case, and we find they often underperform
static A/B tests on practical problem instances with
batched feedback, non-stationarity, multiple objectives and constraints, and
personalization.
In response, we develop a mathematical programming framework for
developing adaptive experimentation algorithms. Instead of the
problem-specific research paradigm (akin to an optimization solver developed
for a particular linear program), we ask the modeler to write down a flexible
optimization formulation and use modern machine learning systems to
(heuristically) solve for adaptive designs.
Since a naive formulation of the adaptive
experimentation problem as a dynamic program is intractable,
we propose a batched view of the experimentation process. We model the uncertainty around
batch-level sufficient
statistics necessary to make allocation decisions, instead of attempting to
model unit-level outcomes whose distributions are commonly unknown and leads
to intractable dynamic programs with combinatorial action spaces.
Sequential Gaussian approximations is the main intellectual vehicle
powering our mathematical programming framework. CLT-based normal approximations are universal in statistical
inference, and a sequential variant we prove provides a simple optimization formulation that lends itself to modern computational tools. Through extensive empirical
evaluation, we observe that even a preliminary and heuristic solution
approach can provide major robustness benefits. Unlike bespoke methods (e.g.,
Thompson sampling variants), our mathematical programming framework provides
consistent gains over static randomized control trials and exhibits robust
performance across problem instances.
@article{CheJiNaWa24,title={Optimization-Driven Adaptive Experimentation},author={Che, Ethan and Jiang, Daniel and Namkoong, Hongseok and Wang, Jimmy},journal={arXiv:2408.04570 [cs.LG]},year={2024},note={Selected for oral presentations at the Econometric Society
Interdisciplinary Frontiers: Economics and AI+ML
conference and Conference on Digital Experimentation},url={https://arxiv.org/abs/2408.04570},}
@article{CaiNaRuZh25,title={Active Exploration via Autoregressive Generation of Missing Data},author={Cai, Tiffany and Namkoong, Hongseok and Russo, Daniel and Zhang, Kelly},journal={arXiv:2405.19466 [cs.LG]},year={2025},note={Selected for presentation at the Econometric Society
Interdisciplinary Frontiers: Economics and AI+ML
conference},url={https://arxiv.org/abs/2405.19466},}
I am unable to respond to most inquiries regarding openings for Ph.D. positions as admissions are handled at the departmental level.