I am a fourth-year PhD student at the NYU Center for Data Science, advised by Kyunghyun Cho and Tal Linzen. I am supported by the NSF Graduate Research Fellowship.
I study the science of training language models, with a focus on data.
- Curriculum learning: [Aioli], [Pre-pretraining]
- Scaling laws: [Neural neural scaling laws], [Downstream scaling laws],
[Scaling laws & modeling humans] - Training dynamics: [Modeling training], [Visualizing training]
Lately, I've started thinking about self-improving models and automating ML research.
Relevant: [Neural neural scaling laws].
Previously, I completed a BSE at Princeton CS, where I spent two lovely years working with Karthik Narasimhan and Tom Griffiths. I then joined Yobi AI for two years as the first employee. In my spare time, I enjoy cooking, running, and playing basketball.