Assistant Professor of Computer Science, Columbia University.
Visiting Researcher, Google DeepMind.
jh5020 [at] columbia.edu
I am a researcher interested in developing neural language systems, deeply understanding them, and precisely controlling them, for the sake of peoples' access to information and useful tools.
Feel free to look me up on Google Scholar or Twitter, or take my CV.
I strive to provide open access to my teaching materials.
You can find materials for my new natural language processing course, a version of Columbia Computer Science 4705.
Older videos of my Stanford 224n lectures are available on YouTube.
Join my lab @ Columbia
Here's an interest form for Columbia undergrads and masters students.
If you're a current Columbia PhD student, please email me.
If you're applying to PhD programs, apply to the Department of Computer Science and include me in the list of professors you're interested in; I plan to admit roughly two students in the upcoming PhD cycle (for students starting Fall 2026.)
More about me
I did my PhD research at Stanford Computer Science, as part of the NLP group. I'm grateful to have been co-advised by Chris Manning and Percy Liang, and to have been supported by an NSF Graduate Research Fellowship.
Before that, I did my undergrad studies at Penn.
Model Editing with Canonical Examples. John Hewitt, Sarah Chen, Lanruo Lora Xi, Edward Adams, Percy Liang, Christopher D. Manning. ArXiv. (pdf)(code)
A non-archival version wonHonorable Mention for Best Paper at the R0-FoMo Workshop at NeurIPS 2023.
Language Probes as V-information Estimators. NLP with Friends. September 9, 2020.
Probing Neural NLP: Ideas and Problems. Berkeley NLP Seminar. November 18, 2019.
Emergent Linguistic Structure in Neural NLP.
Amazon AI. July 25, 2019.
A Structural Probe for Finding Syntax in Word Representations. NLP Highlights Podcast. May, 2019.
Abstracts
RNNs can generate bounded hierarchical languages with optimal memory. John Hewitt, Michael Hahn, Surya Ganguli, Percy Liang, Christopher D. Manning.
2020 Conference on the Mathematical Theory of Deep Learning (abstracts).
Semantic Bootstrapping in Frames: A Computational Model of Syntactic Category Acquisition. John Hewitt, Jordan Kodner, Mitch Marcus, and Charles Yang.
Conference of the Cognitive Science Society (CogSci), (member posters) 2017.
(pdf)(abstract)
Patents
Determining training data sizes for training smaller neural networks using shrinking estimates. John Hewitt, Adhiguna Kuncoro, Aida Nematzadeh.
US Patent App US18/932,554, priority to EP23206829.6A, October 2023. (application)
Capturing Rich Response Relationships with Small-Data Neural Networks. John Hewitt.
US Patent App 15/841,963. December 2017. (granted). (application)
I wrote a lecture on Transformers in my role as Head TA for Stanford's CS 224N: Natural Language Processing with Deep Learning in 2021.
The updated slides are available, as is a recording on YouTube.
In 2023, I updated the lecture (which had also been updated by Anna Goldie in 2022).
Along with the lecture, in 2023 I wrote brand new lecture notes.