excited that the Society for Computation in Linguistics (SCiL) will be colocated with #acl2026nlp this year, and I'm grateful to the National Science Foundation for helping support SCiL presenters' registration costs!
(keynotes: Jenn Hu and Noah Smith
deadline: Jan 30
conference: July 3 & 4)
Posts by Tal Linzen
Really big announcement! See @wtimkey.bsky.social's thread for the details on an exciting new preprint from the NYU-UMass Syntactic Ambiguity Processing group. It is the culmination of the team's research efforts over these last couple of years, and we're really happy with it.
New Preprint: osf.io/eq2ra
Reading feels effortless, but it's actually quite complex under the hood. Most words are easy to process, but some words make us reread or linger. It turns out that LLMs can tell us about why, but only in certain cases... (1/n)
New NeurIPS paper! Why do LMs represent concepts linearly? We focus on LMs's tendency to linearly separate true and false assertions, and provide an analysis of the truth circuit in a toy model. A joint work with Gilad Yehudai, @tallinzen.bsky.social, Joan Bruna and @albertobietti.bsky.social.
๐Introducing BabyBabelLM: A Multilingual Benchmark of Developmentally Plausible Training Data!
LLMs learn from vastly more data than humans ever experience. BabyLM challenges this paradigm by focusing on developmentally plausible data
We extend this effort to 45 new languages!
Another banger from @tallinzen.bsky.social .
Also fits with some of the criticisms of Centaur and my faculty-based approach generally; if you want LLMs to model human cognition, give them more architecture akin to human faculty psychology like long and short-term memory.
arxiv.org/abs/2510.05141
thanks Cameron!
Linguistics PhD student @jacksonpetty.org finds LLMs "quiet-quit" when instructions get long, switching from reasoning to guesswork.
With CDS' @tallinzen.bsky.social, @shauli.bsky.social, @lambdaviking.bsky.social, @michahu.bsky.social, and Wentao Wang.
nyudatascience.medium.com/llms-switch-...
Nature: US senators poised to reject Trumpโs proposed massive science cuts Committee gives first hint that policymakers might preserve, rather than slash, funding for US National Science Foundation and other agencies.
DO NOT GIVE UP!
Our advocacy is working.
A key Senate committee has indicated that it will reject Trumpโs proposed cuts to science agencies including NASA and the NSF.
Keep speaking up and calling your electeds ๐ฃ๏ธ๐ฃ๏ธ๐ฃ๏ธ
Maybe five years with a no-cost extension!
Congratulations to @linguistbrian.bsky.social for receiving this grant to study how to constrain language models to read complex sentences more like humans, and congratulations to me for getting to collaborate with him for another four years! www.umass.edu/humanities-a...
Thanks Andrea!
If we have a lot of shared followers, perhaps you could comment on the pinned tweet on my account and provide context?! Thank you!
My Twitter account has been hacked :( Please don't click on any links "I" posted on that account recently!
I'll be accepting applications for a while, and will also consider people with a late start date. Feel free to email if you have questions. No need for a formal cover letter.
The goal is to model some cool behavioral and neural data from humans (some to be collected) but we expect to do a lot of fundamental modeling and interpretability work. You don't need to have existing experience in cognitive science but you should be interested in learning more about it.
I'm hiring at least one post-doc! We're interested in creating language models that process language more like humans than mainstream LLMs do, through architectural modifications and interpretability-style steering. Express interest here: docs.google.com/forms/d/e/1F...
ืืืืงื ืืืืงืืจืช ืืืจืืื ืื ืืฆื ืืืฉืจืืื ืฉื ืืืขืืจ ืื ืืืงืฉื ืืืืื ืฉืืฉ ืื ืื ืืจืืื ืืจ, ื ืชื ื ืื ืืฆืืช ืืืืจืฅ ืืื ืืขืืื. ืืกืชืืจ ืฉืืฉ ืกืืืื ืืืืื ื ืืืฆืืื ืืืืจืฅ ืฉื ืืืจืืื ืืฉืจืืืื ืจืง ืืืจื ืืืืืืจ, ืื ืืืจื ืืืืฉื
How well can LLMs understand tasks with complex sets of instructions? We investigate through the lens of RELIC: REcognizing (formal) Languages In-Context, finding a significant overhang between what LLMs are able to do theoretically and how well they put this into practice.
Following the success story of BabyBERTa, I & many other NLPers have turned to language acquisition for inspiration. In this new paper we show that using Child-Directed Language as training data is unfortunately *not* beneficial for syntax learning, at least not in the traditional LM training regime
Depends on what you mean by US academics, I guess. A lot of people are here for a temporary position, don't have strong ties to the country, and were mentally prepared to move elsewhere anyway. Those people are much more likely to leave than before.
I'll have a bit of time to chat with folks in Berlin and/or Copenhagen about AI, LLMs, cognitive science, how good your bike infrastructure is, etc, let me know!
And this one on language models with cognitively plausible memory in Potsdam on Tuesday (as part of this in-person-only sentence processing workshop vasishth.github.io/sentproc-wor...):
Cross-posting the abstracts for two talks I'm giving next week! This one on formal languages for LLM pretraining and evaluation, at Apple ML Research in Copenhagen on Wednesday
Updated version of our position piece on how language models can help us understand how people learn and process language, on why it's crucial to train models on cognitive plausible datasets, and on the BabyLM project that addresses this issue.
out of date, should be $300 billion now!
thanks! I'll start with the frens and nice people and work my way up from there!
At #HSP2025, I'll present work with @tallinzen.bsky.social and @shravanvasishth.bsky.social on modeling garden-pathing in a huge benchmark dataset: hsp2025.github.io/abstracts/29.... Statistically decomposing the effect into subprocesses greatly improves predictive fit over just comparing means!
Going to give this website another shot! What are good lists of linguistics, psycholinguistics, NLP and AI accounts?
Thanks Ted for mentioning me in the same tweet as Chris! This website really is better than the other one!