Congrats to Pingjun, @beiduo.bsky.social , Siyao, Marie, and @barbaraplank.bsky.social for receiving the SAC Highlights reward!
Posts by Beiduo Chen
What an incredible EMNLP experience — truly the most fulfilling conference I’ve ever attended!
✅ Oral presentation
✅ SAC Highlights Award
✅ Panel discussion
Grateful to my amazing collaborators and to all the friends I had the chance to meet! 🌟
#EMNLP2025 #NLP
Detailed programme now up on website. Looking forward to 14 research papers, results of the 3rd Shared Task on Learning with Disagreements (LeWiDi), a talk from @camachocollados.bsky.social, and a panel discussion feat. Jose, Eve Fleisig, and @beiduo.bsky.social. See you in Room A305 or online!
Our paper: arxiv.org/pdf/2505.23368
Our code: github.com/mainlp/CoT2EL
Thank you to my wonderful co-authors,
@janetlauyeung.bsky.social, Anna Korhonen, and @barbaraplank.bsky.social. Also to @mainlp.bsky.social , @cislmu.bsky.social @munichcenterml.bsky.social
See you in Suzhou!
#NLP #EMNLP2025
Matching exact probabilities for HLV is unstable. So, we propose a more robust rank-based evaluation that checks preference order. Our combined method outperforms baselines on 3 datasets that exhibit human label variation, showing it better aligns with diverse human perspectives.
Instead of unnatural post-hoc explanations, we look forward. A model's CoT already contains rationales for all options. We introduce CoT2EL, a pipeline that uses linguistic discourse segmenters to extract these high-quality, faithful units to explore human label variation.
📑 Our CoT2EL paper will be presented as an oral at #EMNLP2025 in Suzhou!
Humans often disagree on labels. Can a model's own reasoning (CoT) help us understand why? We developed a new method to extract these insights. Come join us!
🗓️ Friday, Nov 7, 14:00 - 15:30
📍 Room: A110
🌍 Broader impact:
Our approach makes capturing disagreement scalable, helping build datasets that reflect real-world ambiguity—without requiring tons of human-written explanations.
Open-sourcing:
📂 github.com/mainlp/MJD-E...
🧠 What’s this about?
Human annotations often disagree. Instead of collapsing disagreement into a single label, we model Human Judgment Distributions — how likely humans are to choose each label in NLI tasks.
Capturing this is crucial for interpretability and uncertainty in NLP.
🔗 Paper link: arxiv.org/abs/2412.13942
🙏 Huge thanks to our collaborators Logan Siyao Peng, @barbaraplank.bsky.social, Anna Korhonen from @mainlp.bsky.social, @lmumuenchen.bsky.social, @cambridgeltl.bsky.social
🚨 Can LLMs generate explanations that are as useful as human ones for modeling label distributions in NLI?🌹"A Rose by Any Other Name" shows that they can
💬 We explore scalable, explanation-based annotation via LLMs.
📍Come find us in Vienna 🇦🇹! (July 28, 18:00-19:30, Hall 4/5) #ACL2025NLP #acl2025
The hand-drawn sign from three years ago.
🎉MaiNLP is turning 3 today!🎂🥳 We’ve grown a lot since @barbaraplank.bsky.social started this group with nothing but three aspiring researches and a hand-drawn sign on the door. Huge thanks to all the amazing people who have joined or visited us since. Here’s to many more years of exciting research!🚀