Michael Hu

IMG_1151-EDIT.jpg

I am a third-year PhD student at the NYU Center for Data Science, working with Kyunghyun Cho and Tal Linzen. I work on algorithms that optimize the training data of language models. I am supported by an NSF Graduate Research Fellowship.

In my spare time, I enjoy cooking, running, and playing basketball.

Previously, I completed a BSE at Princeton CS, where I spent two lovely years working with Karthik Narasimhan and Tom Griffiths.

news

selected publications

2025

  1. Preprint
    Between Circuits and Chomsky: Pre-pretraining on Formal Languages Imparts Linguistic Biases
    Michael Y. Hu, Jackson Petty, Chuan Shi, and 2 more authors
    Preprint, 2025
  2. ICLR
    Aioli: A Unified Optimization Framework for Language Model Data Mixing
    Mayee F. Chen*, Michael Y. Hu*, Nicholas Lourie, and 2 more authors
    ICLR, 2025

2024

  1. Preprint
    Bigger is not always better: The importance of human-scale language modeling for psycholinguistics
    Ethan Wilcox, Michael Y. Hu, Aaron Mueller, and 6 more authors
    Preprint, 2024

2023

  1. TMLR
    Latent State Models of Training Dynamics
    Michael Y. Hu, Angelica Chen, Naomi Saphra, and 1 more author
    Transactions on Machine Learning Research, 2023

2021

  1. NeurIPS
    Safe Reinforcement Learning with Natural Language Constraints
    Tsung-Yen Yang, Michael Hu, Yinlam Chow, and 2 more authors
    NeurIPS, 2021