thanks, Robert!!
Posts by Haoran Zhao
Had an amazing experience at EMNLP 2025 @emnlpmeeting.bsky.social. Glad to present my work in an oral session and honored to win the "SAC Highlight" award. Feel free to check the work below. Big thanks to my amazing advisor @rdhawkins.bsky.social!
π€ "Your #CogSci presentation was quite good this year."
How flattered or offended will you be? The answer may depend on whether you speak British or American English πΊπΈπ¬π§. Our new #CogSci2025 paper reveals systematic differences in how different cultures interpret the same words.
Work done with the amazing @rdhawkins.bsky.social. Looking forward to presenting on this work at CogSci 2025 this summer. Please check our full paper at: arxiv.org/abs/2506.09391
In summary, we find that while LLMs have impressive politeness capabilities, their systematic preference for distancing strategies reveals important gaps in pragmatic alignment. Future work should explore how to better balance positive and negative politeness strategies. π―
As we deploy LLMs in social contexts, we need to think beyond whether they CAN be polite to HOW they're polite. Training that emphasizes "harmlessness" may inadvertently create systems that are pragmatically misaligned with human communication patterns.
Why does this matter? Despite good agreement on multi-choice tasks, subtle misalignments in open-ended polite language production could lead to real communication breakdowns: Hedged positive feedback might be interpreted as more negative than the system intends to convey.
Result #3: LLMs systematically overuse negative politeness strategies in positive contexts! While humans shift to rapport-building language for good performances, LLMs keep hedging and distancing.
So LLMs are "better" at politeness? Not necessarily. A deeper dive reveals a crucial difference. Politeness theory distinguishes between positive strategies, which build rapport ("I love your creativity here!"), and negative strategies, which minimize imposition ("I'm somewhat concerned...").
Result #2: When we removed constraints and let both humans and LLMs freely generate responses, human evaluators actually PREFERRED the LLM responses 66% of the time! π€― This held across all communicative goals!
But real conversations aren't multiple choice questions! So we ran a bigger test: What happens when humans and LLMs can say anything they want? We collected XXX+ responses for the same scenarios, manipulating the goal (being informative, kind, or both).
Result #1: Models β₯70B successfully replicated human patterns (like using 'wasn't terrible' for a bad performance). Smaller models could barely do the task properly. π
We started by asking how LLMs compare to human politeness preferences in a simple task introduced by Yoon et al (2020). If a friend gives a bad performance (0/3 β€οΈ), what would you say to them? Humans use negation to soften the blow: 'it wasn't terrible' is preferred over "it was bad."
LLMs are increasingly deployed in sensitive social contexts like education, healthcare, and customer service. If they're systematically different in HOW they're polite, it could lead to misunderstandings.
Suppose your friend asks 'How was my cooking?' and it was... not great. π¬ Speakers use complex politeness strategies to navigate tricky situations. But what happens when LLMs do? Weβre excited to share new work revealing surprising similarities and differences in human and LLM politeness usageπ§΅.