today at 12:10!
Posts by Anzi Wang
Friday time! Poster 5 (# 365) @grushaprasad.bsky.social , @shotamomma.bsky.social and I leverage an ACT-R based parsing model (from Grusha’s previous work!) to evaluate potential explanations of gradient effects in cross-structural priming. Explicit parsers FTW!!
TunePad (tunepad.com)! Developed by amazing researchers at Northwestern, extremely friendly and intuitive UI. I'm teaching fifth-graders at local schools TunePad and most of them are loving it
one year ago i thought p-side was about predicates and s-side was about subjects :3
Wow!! Congratulations!!
Reposting because the link has expired:
PDF: drive.google.com/file/d/1t2EF... (if this doesn't work, lmk)
Publisher link: www.sciencedirect.com/science/arti...
so cool!!
New Preprint: osf.io/eq2ra
Reading feels effortless, but it's actually quite complex under the hood. Most words are easy to process, but some words make us reread or linger. It turns out that LLMs can tell us about why, but only in certain cases... (1/n)
Screenshot of a figure with two panels, labeled (a) and (b). The caption reads: "Figure 1: (a) Illustration of messages (left) and strings (right) in toy domain. Blue = grammatical strings. Red = ungrammatical strings. (b) Surprisal (negative log probability) assigned to toy strings by GPT-2."
New work to appear @ TACL!
Language models (LMs) are remarkably good at generating novel well-formed sentences, leading to claims that they have mastered grammar.
Yet they often assign higher probability to ungrammatical strings than to grammatical strings.
How can both things be true? 🧵👇
I took Grusha and Forrest's version of NLP at Colgate (arxiv.org/abs/2408.05664) and as a current linguistics phd student still doing NLP research, I can say that this is THE undergrad course that has benefited me the most
I've found it kind of a pain to work with resources like VerbNet, FrameNet, PropBank (frame files), and WordNet using existing tools. Maybe you have too. Here's a little package that handles data management, loading, and cross-referencing via either a CLI or a python API.
Brand new version of this paper (now a short book!) available at lingbuzz.net/lingbuzz/008...!
favorite garden path sentence of the year: "It's better to be hurt by someone you know accidentally, than by a stranger on purpose" by Dwight Schrute
probably? maybe you can have different policies each semester and test it out lol
might depend more on participation policy
A paper with Vic Ferreira and Norvin Richards is now out
(1) Speakers syntactically encode zero complementizers as cognitively active mental object.
(2) No evidence LLMs capture cross constructional generalizations about null complementizers.
nam10.safelinks.protection.outlook.com?url=https%3A...
Work with kittens? Check out the National Kitten Coalition's new Kitten Resource Library! They're an org I like a lot!! library.kittencoalition.org
"We introduce Collaborative Rational Speech Act (CRSA), an information-theoretic (IT) extension of RSA that models multi-turn dialog by optimizing a gain function adapted from rate-distortion theory."
arxiv.org/abs/2507.14063
Someone asked me today how to get better at scientific writing. I'm not the best person to ask because I find my own writing very inadequate! But the tips I thought of were:
1. Practice, and practice with co-authors who are better writers than you. Observe how they make edits and copy them.
(1/n)
early November is also the best season for crabs!!
is computational psycholinguistics a poly-sci? #puns #linguistics