Advertisement ยท 728 ร— 90

Posts by Elinor

Preview
Jacob Andreas and Brett McGuire named Edgerton Award winners MIT associate professors Jacob Andreas and Brett McGuire have been selected as the winners of the 2026ย Harold E. Edgerton Faculty Achievement Award for exceptional contributions to teaching, research,...

Congratulations to Jacob Andreas, was named a 2026 Edgerton Award recipient!
The award recognizes exceptional teaching, research, and service at MIT! Prof. Andreas co-leads our Language and Thought Mission, and he is a dedicated and creative researcher and educator.
news.mit.edu/2026/jacob-a...

4 days ago 7 1 1 0
Post image Post image

My Hoya Rebecca put out the cutest teeny lil flowers today and Iโ€™m obsessed

3 days ago 3 0 0 0

Tech industry mottos have a mixed track record. But we should hold idealists to their ideals. And we should celebrate when they come through.

The Mythos non-release is a remarkable moment of conviction. Thoughts:
davidbau.com/archives/20...

Bravo to Anthropic's "race the top".

1 week ago 13 3 1 0
Preview
Building AI for the Democratic Matrix: A Technical Research Agenda for Normative Competence and Normative Institutions To maintain democratic resilience, it is essential to build AI agents capable of choosing behaviors that mirror those of the human agents that constitute human democracies.

Democracy isn't a rulebook. It runs on daily interactions where people comply with norms and hold each other accountable. AI agents are about to join that system. We need to build them to read it. New paper with Rakshit Trivedi and Dylan Hadfield-Menell.

2 weeks ago 11 2 4 0
Post image

Rebuttal season is here, yey๐Ÿคž

With many asking me,
I compiled the most common misconceptions

Hope the tips help ๐Ÿงต

All tips:
docs.google.com/document/d/14Wax8M5w8F_8miDlYJ9-I6wqpelxlXjCEUbkNzNMqqE/edit?tab=t.0#heading=h.rfq27f356vmm
#AI
๐Ÿค–๐Ÿ“ˆ๐Ÿง 

3 weeks ago 5 2 1 0

New important (I hope) resource for academics working in this area.

4 weeks ago 9 1 1 0

For a recent lab meeting, I wrote up a grab bag of ways to think about your development as a researcher during a PhD: emerge-lab.github.io/papers/an-un...

Sharing in case folks find it useful or have feedback!

1 month ago 102 12 6 5

its worthwhile for future work to actually understand which types of queries we want Overton pluralistic responses (& how it relates to human prefs)

Using the metric to understand models pluralistic capabilities vs directly as a reward/optimization are different, its not designed for the latter rn

1 month ago 2 0 0 0
Advertisement

The paper frames higher as better but doesnโ€™t assert that all models should strive for perfect scores all the time. And the *subjective* queries on which we want any Overton pluralism is pretty narrow.

The past example is rlly interesting though! Hints at a general problem of Overton Window shifts

1 month ago 1 0 1 0
Video

๐Ÿฅ๐Ÿฅ๐Ÿฅ Newly out from us today in Science Advances: โ€œBiased AI Writing Assistants Shift Usersโ€™ Attitudes on Societal Issuesโ€.

Large Language Models are providing users with autocomplete writing suggestions on many platforms. Could these suggestions shift usersโ€™ own attitudes? (spoiler: YES) (1/7)

1 month ago 188 104 4 19

the key points are

1) its important to measure pluralistic capability bc we think its necessary for better overall value alignment
2) its especially important to understand its impacts on users (future work) + tradeoffs wrt political neutrality, etc

1 month ago 1 0 1 0

i dont think so, & thats not what the paper is really about imo :)

strictly perfect overton pluralism isn't the actual goal for model behavior across the board. however, models struggle on many subjective qs + improving overton pluralism for those types of responses is important

1 month ago 0 0 1 0

Huge thanks to my amazing coauthors ๐Ÿ™ Jiayi Wu @taylor-sorensen.bsky.social Jiaxin Pei @mbakker.bsky.social !

Excited to keep pushing on pluralistic alignment. Please reach out if you want to connect ๐Ÿ’ฌ๐Ÿค—

Paper: arxiv.org/abs/2512.01351
Website: overtonbench.github.io

9/9

1 month ago 1 0 0 0

Inspired by
@bennokrojer.bsky.social, we included a Behind the Scenes section ๐ŸŽฌ

The goal is to make science more transparent ๐Ÿ”, share lessons learned ๐Ÿง , and provide a more realistic lens on the research journey ๐Ÿ‘ฃ

8/

bsky.app/profile/benn...

1 month ago 5 1 1 0

However, human studies aren't scalable๐Ÿ’ฐ

We build + validate an LLM-as-judge that approximates human representation scores so you can use ๐Ž๐•๐„๐‘๐“๐Ž๐๐๐„๐๐‚๐‡ without running a new study each time

We open-source our code to foster development of more pluralistic LLMs ๐Ÿš€

7/

1 month ago 0 0 1 0
Advertisement
Post image

A key finding: neutral โ‰  pluralistic

A politically balanced or neutral response can still fail to represent large swaths of viewpoints

We find political slant and pluralism are ๐™ฃ๐™š๐™œ๐™–๐™ฉ๐™ž๐™ซ๐™š๐™ก๐™ฎ ๐™˜๐™ค๐™ง๐™ง๐™š๐™ก๐™–๐™ฉ๐™š๐™™ and ๐™™๐™ž๐™จ๐™ฉ๐™ž๐™ฃ๐™˜๐™ฉ concepts

6/

1 month ago 3 1 1 0
Post image Post image Post image

So how do current models do? ๐Ÿ‘€

Best-performing models score 0.35โ€“0.41 well below 1 (max)

A lot of room to grow โ€” and we discuss in the paper interesting variation across models and topics, pointing to where alignment efforts should focus

5/

1 month ago 0 0 1 0
Polis

To determine ๐™™๐™ž๐™จ๐™ฉ๐™ž๐™ฃ๐™˜๐™ฉ viewpoints, we ran a 1,200+ person US-representative human study ๐Ÿง‘โ€๐Ÿคโ€๐Ÿง‘and cluster

๐Ÿ’กKey: instead of algorithmic clustering, users vote to group themselves, inspired by pol.is + is more faithful to the underlying perspectives

4/

1 month ago 0 0 1 0
Post image

To operationalize, we introduce a set-coverage metric

For each question, we calculate the proportion of ๐™™๐™ž๐™จ๐™ฉ๐™ž๐™ฃ๐™˜๐™ฉ viewpoints ๐Ÿ—ฃ๏ธ covered by each model response.

We determine coverage by directly asking humans whether their POV is represented in the model response

3/

1 month ago 0 0 1 0
Post image

๐Ž๐•๐„๐‘๐“๐Ž๐๐๐„๐๐‚๐‡ measures Overton pluralism:

For a subjective query, to what extent does a model's response represent the โœจfullโœจ range of reasonable viewpoints?

2/

1 month ago 0 0 1 0
Post image

There's been a lot of excitement about pluralistic value alignment ๐ŸŒˆ โ€” AI that reflects the full range of human perspectives

But no formal way to benchmark whether we're actually making progress. ๐Ÿค”

Introducing ๐Ž๐•๐„๐‘๐“๐Ž๐๐๐„๐๐‚๐‡. ๐ŸŽ‰Accepted to #ICLR2026

1/n ๐Ÿงต

1 month ago 20 1 1 1
Post image

Do LLMs Benefit from Their Own Words?๐Ÿค”

In multi-turn chats, models are typically given their own past responses as context.
But do their own words always helpโ€ฆ
Or are they more often a waste of compute and a distraction?
๐Ÿงต
arxiv.org/abs/2602.24287

1 month ago 37 4 2 1
Title, author list, and two figures from the paper. 
Title: The Aftermath of DrawEduMath: Vision Language Models
Underperform with Struggling Students and Misdiagnose Errors
Authors: Li Lucy, Albert Zhang, Nathan Anderson, Ryan Knight, Kyle Lo
Figure 1: On the left is a math problem, where students are asked to draw x < 5/2 on a number line. The right side shows two example student responses that differ in correctness. DrawEduMath pairs each math problem with one student response, and prompts VLMs to answer questions about the student response.
Figure 2: VLMs consistently perform worse on answering DrawEduMath benchmark questions pertaining to erroneous student responses. Performance on non-erroneous student responses is labeled with specific VLMsโ€™ names; that same modelโ€™s performance on erroneous student responses is directly below.

Title, author list, and two figures from the paper. Title: The Aftermath of DrawEduMath: Vision Language Models Underperform with Struggling Students and Misdiagnose Errors Authors: Li Lucy, Albert Zhang, Nathan Anderson, Ryan Knight, Kyle Lo Figure 1: On the left is a math problem, where students are asked to draw x < 5/2 on a number line. The right side shows two example student responses that differ in correctness. DrawEduMath pairs each math problem with one student response, and prompts VLMs to answer questions about the student response. Figure 2: VLMs consistently perform worse on answering DrawEduMath benchmark questions pertaining to erroneous student responses. Performance on non-erroneous student responses is labeled with specific VLMsโ€™ names; that same modelโ€™s performance on erroneous student responses is directly below.

Models are now expert math solvers, and so AI for math education is receiving increasing attention.
Our new preprint evaluates 11 VLMs on our QA benchmark, DrawEduMath. We highlight a startling gap: models perform less well on inputs from K-12 students who need more help. ๐Ÿงต

1 month ago 36 12 4 2
Advertisement

Yesterday was my last day at MSR. We recently learned that our roles were eliminated, and with them our little FATE Montreal team.

I joined MSR a bit over 7.5 years ago while on active chemotherapy, and being at MSR has overlapped with so much change in my life.

1 month ago 34 6 4 0
Post image

Our paper, "What's in My Human Feedback", received an oral presentation at ICLR!

Our method automatically+interpretably identifies preferences in human feedback data; we use this to improve personalization + safety.

Reach out if you have data/use cases to apply this to!

arxiv.org/pdf/2510.26202

1 month ago 28 3 0 0
Post image

Finally we do test it empirically: finding some models where the embedding matrix of the LLM already provides decently interpretable nearest neighbors

But this was not the full story yet...
@mariusmosbach.bsky.social and @elinorpd.bsky.social nudged me to use contextual embeddings

2 months ago 1 1 1 0

Really cool new work with surprising results! Highly recommend checking out the demo ๐Ÿ‘€

2 months ago 3 0 0 0

Grok fact-checks our paper on Grok fact-checking - and it approves!

2 months ago 28 7 1 0
Post image

๐ŸŽญ How do LLMs (mis)represent culture?
๐Ÿงฎ How often?
๐Ÿง  Misrepresentations = missing knowledge? spoiler: NO!

At #CHI2026 we are bringing โœจTALESโœจ a participatory evaluation of cultural (mis)reps & knowledge in multilingual LLM-stories for India

๐Ÿ“œ arxiv.org/abs/2511.21322

1/10

2 months ago 46 22 1 2

this is amazing! made quick NYC & boston posters

2 months ago 3 0 0 0