senior staff research scientist

at Google DeepMind

beenkim at csail dot mit dot edu
I am working to ensure that as machines get smarter, humans do too.

My work empowers humans to maintain agency by extracting and teaching useful knowledge from AI to humans. We get AI involved in this teaching (agentic interpretability) and this may involve new language (neologism). We proved that this is possible; using AlphaZero knowledge to teach grandmasters, one of whom became the youngest World Chess Champion (PNAS).
Quanta magazine (written by John Pavlus) is a great description of what I do and why.


I gave keynote at ICLR 2022, ECML 2020 and at the G20 meeting in Argentina in 2018. One of my work TCAV received UNESCO Netexplo award, was featured at Google I/O 19' and in Brian Christian's book on The Alignment Problem.

Stuff I help with:
        ICLR board
        General Chair at ICLR 2024
        Senior Program Chair at ICLR 2023
        Steering committee SaTML
        Workshop Chair at ICLR 2019
        Senior Area chair/Area chair/Senior program committee NeurIPS 2017-now, ICML 2019-now, ICLR 2020-now, AISTATS 2020-now
        Steering committee and area chair at FAccT conference
        Former executive board member and VP of Women in Machine Learning.
        Co-organizer of multi-year workshops of Human interpretability in ML (WHI) at ICML 2020 2018 2017 2016 , and NIPS 2016 Worshop on Interpretable Machine Learning for Complex Systems.


Tutorials on interpretability:

        Machine Learning Summer School MLSS 2021 (slides, video)
        Deep Learning Summer school at University of Toronto, Vector institute in 2018 (slides, video)
        CVPR 2018 (slides and videos)
        Tutorial on Interpretable machine learning at ICML 2017 (slides, video).



Blogs

Images
ICLR 2022 keynote
talk video
blog post (covers only the intro part of the talk)