Advertisement · 728 × 90

Posts by Myra Cheng @ ICLR :)

happy to chat about sycophancy, public perceptions of AI, AI interaction through a pragmatic lens, anthropomorphism, or anything else :-) please reach out!

Assumptions preprint: arxiv.org/pdf/2604.03058

1 week ago 2 0 0 0

In Barcelona for #chi2026! Presenting our work on eliciting LLMs' assumptions about users, and how this mismatches with user expectations, in the Tues poster session! (Spoiler: users assume that LLMs give objective info much more than they actually do, which leads to sycophancy)

1 week ago 15 1 1 0

Many are appropriately outraged by Altman’s comments here implying that raising a human child is akin to “training” an AI model.

This is part of a broader pattern where AI industry leaders use language that collapses the boundary between human and machine.

🧵/

1 month ago 493 200 28 22
Photo of Cornelll University building surrounded by colorful trees

Photo of Cornelll University building surrounded by colorful trees

No better time to start learning about that #AI thing everyone's talking about...

📢 I'm recruiting PhD students in Computer Science or Information Science @cornellbowers.bsky.social!

If you're interested, apply to either department (yes, either program!) and list me as a potential advisor!

5 months ago 23 9 1 0
A circular flow diagram that compares current and proposed practices for LLM development using data from adopters and non-adopters. Three gray boxes represent current practices: “R&D,” “Chat Models,” and “Adopters’ Needs and Usage Data,” connected in a clockwise loop with black arrows. A blue box labeled “Non-adopters’ Needs and Usage Data” adds a proposed feedback path, shown with blue arrows, linking non-adopter data back to R&D and adopters’ data.

A circular flow diagram that compares current and proposed practices for LLM development using data from adopters and non-adopters. Three gray boxes represent current practices: “R&D,” “Chat Models,” and “Adopters’ Needs and Usage Data,” connected in a clockwise loop with black arrows. A blue box labeled “Non-adopters’ Needs and Usage Data” adds a proposed feedback path, shown with blue arrows, linking non-adopter data back to R&D and adopters’ data.

As of June 2025, 66% of Americans have never used ChatGPT.

Our new position paper, Attention to Non-Adopters, explores why this matters: AI research is being shaped around adopters—leaving non-adopters’ needs, and key LLM research opportunities, behind.

arxiv.org/abs/2510.15951

5 months ago 38 13 2 0

I'll be at COLM next week! Let me know if you want to chat! @colmweb.org

@neilrathi.bsky.social will be presenting our work on multilingual overconfidence in language models and the effects on human overreliance!

arxiv.org/pdf/2507.06306

6 months ago 7 1 0 0
Abstract and results summary

Abstract and results summary

🚨 New preprint 🚨

Across 3 experiments (n = 3,285), we found that interacting with sycophantic (or overly agreeable) AI chatbots entrenched attitudes and led to inflated self-perceptions.

Yet, people preferred sycophantic chatbots and viewed them as unbiased!

osf.io/preprints/ps...

Thread 🧵

6 months ago 177 91 5 15
Advertisement

Was a blast working on this with @cinoolee.bsky.social @pranavkhadpe.bsky.social, Sunny Yu, Dyllan Han, and @jurafsky.bsky.social !!! So lucky to work with this wonderful interdisciplinary team!!💖✨

6 months ago 1 0 0 0

While our work focuses on interpersonal advice-seeking, concurrent work by @steverathje.bsky.social @jayvanbavel.bsky.social
et al. finds similar patterns for political topics, where sycophantic AI also led to more extreme attitudes when users discussed gun control, healthcare, immigration, etc.!

6 months ago 2 0 1 0

There is currently little incentive for developers to reduce sycophancy. Our work is a call to action: we need to learn from the social media era and actively consider long-term wellbeing in AI development and deployment. Read our preprint: arxiv.org/pdf/2510.01395

6 months ago 8 1 1 0
Rightness judgment is higher and repair likelihood is lower for sycophantic AI

Rightness judgment is higher and repair likelihood is lower for sycophantic AI

Response quality, return likelihood, and trust are higher for sycophantic AI

Response quality, return likelihood, and trust are higher for sycophantic AI

Despite sycophantic AI’s reduction of prosocial intentions, people also preferred it and trusted it more. This reveals a tension: AI is rewarded for telling us what we want to hear (immediate user satisfaction), even when it may harm our relationships.

6 months ago 8 3 1 0
Description of Study 2 (hypothetical vignettes) and Study 3 (live interaction) where self-attributed wrongness and desire to initiate repair decrease, while response quality and trust increases.

Description of Study 2 (hypothetical vignettes) and Study 3 (live interaction) where self-attributed wrongness and desire to initiate repair decrease, while response quality and trust increases.

Next, we tested the effects of sycophancy. We find that even a single interaction with sycophantic AI increased users’ conviction that they were right and reduced their willingness to apologize. This held both in controlled, hypothetical vignettes and live conversations about real conflicts.

6 months ago 8 3 1 2
Description of Study 1, where we characterize the prevalence of social sycophancy and find it to be highly prevalent across leading AI models

Description of Study 1, where we characterize the prevalence of social sycophancy and find it to be highly prevalent across leading AI models

We focus on the prevalence and harms of one dimension of sycophancy: AI models endorsing users’ behaviors. Across 11 AI models, AI affirms users’ actions about 50% more than humans do, including when users describe harmful behaviors like deception or manipulation.

6 months ago 6 0 1 0
Screenshot of paper title: Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence

Screenshot of paper title: Sycophantic AI Decreases Prosocial Intentions and Promotes Dependence

AI always calling your ideas “fantastic” can feel inauthentic, but what are sycophancy’s deeper harms? We find that in the common use case of seeking AI advice on interpersonal situations—specifically conflicts—sycophancy makes people feel more right & less willing to apologize.

6 months ago 115 48 2 7

Thoughtful NPR piece about ChatGPT relationship advice! Thanks for mentioning our research :)

8 months ago 12 0 0 0

Congrats Maria!! All the best!!

8 months ago 3 0 0 0

#acl2025 I think there is plenty of evidence for the risks of anthropomorphic AI behavior and design (re: keynote) -- find @myra.bsky.social and I if you want to chat more about this or our "Dehumanizing Machines" ACL 2025 paper

8 months ago 11 1 0 0
Advertisement
Preview
Computer-vision research powers surveillance technology - Nature An analysis of research papers and citing patents indicates the extensive ties between computer-vision research and surveillance.

New paper hot off the press www.nature.com/articles/s41...

We analysed over 40,000 computer vision papers from CVPR (the longest standing CV conf) & associated patents tracing pathways from research to application. We found that 90% of papers & 86% of downstream patents power surveillance

1/

9 months ago 954 532 34 77

Aw thanks!! :)

9 months ago 1 0 0 0

Paper: arxiv.org/pdf/2502.13259
Code: github.com/myracheng/hu...
Thanks to my wonderful collaborators Sunny Yu and @jurafsky.bsky.social and everyone who helped along the way!!

10 months ago 0 0 2 0
Plots showing that DumT reduces MeanHumT and has higher performance on RewardBench than the baseline models.

Plots showing that DumT reduces MeanHumT and has higher performance on RewardBench than the baseline models.

So we built DumT, a method using DPO + HumT to steer models to be less human-like without hurting performance. Annotators preferred DumT outputs for being: 1) more informative and less wordy (no extra “Happy to help!”) 2) less deceptive and more authentic to LLMs’ capabilities.

10 months ago 2 0 1 1
human-like LLM outputs are strongly positively correlated with social closeness, femininity, and warmth (r = 0.87, 0.47, 0.45), and strongly negatively correlated with status (r = 0.80).

human-like LLM outputs are strongly positively correlated with social closeness, femininity, and warmth (r = 0.87, 0.47, 0.45), and strongly negatively correlated with status (r = 0.80).

We also develop metrics for implicit social perceptions in language, and find that human-like LLM outputs correlate with perceptions linked to harms: warmth and closeness (→ overreliance), and low status and femininity (→ harmful stereotypes).

10 months ago 1 0 2 0
bar plot showing that human-likeness is lower in preferred responses

bar plot showing that human-likeness is lower in preferred responses

First, we introduce HumT (Human-like Tone), a metric for how human-like a text is, based on relative LM probabilities. Measuring HumT across 5 preference datasets, we find that preferred outputs are consistently less human-like.

10 months ago 3 1 1 0
Screenshot of first page of the paper HumT DumT: Measuring and controlling human-like language in LLMs

Screenshot of first page of the paper HumT DumT: Measuring and controlling human-like language in LLMs

Do people actually like human-like LLMs? In our #ACL2025 paper HumT DumT, we find a kind of uncanny valley effect: users dislike LLM outputs that are *too human-like*. We thus develop methods to reduce human-likeness without sacrificing performance.

10 months ago 23 6 1 0

thanks!! looking forward to seeing your submission as well :D

10 months ago 1 0 0 0
Advertisement

thanks Rob!!

10 months ago 0 0 0 0
Preview
GitHub - myracheng/elephant Contribute to myracheng/elephant development by creating an account on GitHub.

We also apply ELEPHANT to identify sources of sycophancy (in preference datasets) and explore mitigations. Our work enables measuring social sycophancy to prevent harms before they happen.
Preprint: arxiv.org/abs/2505.13995
Code: github.com/myracheng/el...

10 months ago 3 0 0 0
Preview
Social Sycophancy: A Broader Understanding of LLM Sycophancy A serious risk to the safety and utility of LLMs is sycophancy, i.e., excessive agreement with and flattery of the user. Yet existing work focuses on only one aspect of sycophancy: agreement with user...

Oops, yes! arxiv.org/abs/2505.13995

10 months ago 3 0 0 0

Grateful to work with Sunny Yu (undergrad!!!) @cinoolee.bsky.social @pranavkhadpe.bsky.social @lujain.bsky.social @jurafsky.bsky.social on this! Lots of great cross-disciplinary insights:)

10 months ago 7 0 1 0

We also apply ELEPHANT to identify sources of sycophancy (in preference datasets) and explore mitigations. Our work enables measuring social sycophancy to prevent harms before they happen.
Preprint: arxiv.org/abs/2505.13995
Code: github.com/myracheng/el...

10 months ago 2 0 0 0