Michael Hu

IMG_1151-EDIT.jpg

I am a fourth-year PhD student at the NYU Center for Data Science, advised by Kyunghyun Cho and Tal Linzen. I am supported by the NSF GRFP.

I study how to train and adapt large language models.

I’m also interested in ML x {cognitive science, visualization}.

Previously, I completed a BSE at Princeton CS, where I spent two lovely years working with Karthik Narasimhan and Tom Griffiths. I then joined Yobi AI for two years as the first employee.

In my spare time, I enjoy cooking, running, and playing basketball.

news

selected publications

2025

  1. ACL
    Between Circuits and Chomsky: Pre-pretraining on Formal Languages Imparts Linguistic Biases
    Michael Y. Hu, Jackson Petty, Chuan Shi, and 2 more authors
    In Proceedings of the Association for Computational Linguistics, Jul 2025
  2. ICLR
    Aioli: A Unified Optimization Framework for Language Model Data Mixing
    Mayee F. Chen*, Michael Y. Hu*, Nicholas Lourie, and 2 more authors
    ICLR, Jul 2025
  3. JML
    Bigger is not always better: The importance of human-scale language modeling for psycholinguistics
    Ethan Wilcox, Michael Y. Hu, Aaron Mueller, and 6 more authors
    Journal of Memory and Language, Jul 2025

2023

  1. TMLR
    Latent State Models of Training Dynamics
    Michael Y. Hu, Angelica Chen, Naomi Saphra, and 1 more author
    Transactions on Machine Learning Research, Jul 2023