We are delighted to welcome @marlutz.bsky.social to our lab over the next few months! 🎉
She'll work on the representation of different demographic groups in LLMs.
#NLProc
Posts by Martin Tutek
FYI #ACL2026 has an unusual registration system this year, and probably a lot of people who want to attend will not be able to.
Spots are limited to 3.5k people, and only presenting authors can register during the first phase. Then, *if* there are spots left, others can try to register.
📢 The workshop on Insights from negative results will be back at EMNLP'26!
Your most-insightful failures can be submitted in 4 pages by June 25. It's also possible to commit short papers reviewed through ARR.
insights-workshop.github.io/2026/cfp
How can generative AI better support human creativity, without limiting it? If you have thoughts, we invite submissions to our ICML workshop on Generative AI, Creativity, and Human-AI Co-Creation
📍 July 2026, Seoul
📄 Submit by: April 24 (AOE)
🔗 Submission link: openreview.net/group?id=ICM...
A llama sweating while writing a paper at a desk. A sign says "Deadline! March 31 11:59pm AOE"
❗The full paper submission deadline for COLM is ~14 hours from now (11:59pm AOE)!
Please submit your final PDFs on the same page where you uploaded your abstracts. And please use the provided LaTeX templates; do not handwrite your manuscript like this llama is!
Good luck!
Interested in pursuing a PhD in NLP/cog-sci?
Studying language learning in LMs from the perspective of human language acquisition? Few more days to apply!!
I notice a surprising lack of emdashes in this post, do you not like them?
A piece co-authored by an old friend (Divya Saini, a psychiatrist at Massachusetts General Hospital)
www.nytimes.com/2026/03/29/o...
Check out works on sequence repetition 🔁 and evaluating synthetic data 🧮 from our lab in Rabat!
@eaclmeeting.bsky.social #EACL2026
The Curse of Verbalization: How Presentation Order Constrains LLM Reasoning
aclanthology.org/2026.finding...
> Restructuring problems to align the order of information presentation with the order of utilization consistently improves performance.
Intuitive but neat.
Mary, the Cheeseburger-Eating Vegetarian: Do LLMs Recognize Incoherence in Narratives?
aclanthology.org/2026.eacl-lo...
How LLMs process contradictory information in context (wrt. what they expect or know) is a big question. Contradictions in settings seem more important to traits.
Rethinking Hallucinations: Correctness, Consistency, and Prompt Multiplicity
aclanthology.org/2026.eacl-lo...
To put it plainly, hallucinations are a frustratingly poorly defined phenomenon. To mitigate it, nuance and categorization is important, which this paper does a good job of.
LLMs Faithfully and Iteratively Compute Answers During CoT: A Systematic Analysis With Multi-step Arithmetics
aclanthology.org/2026.finding...
Big fan of faithfulness work, causal interventions and even moreso in specialized scenarios (arithmetic) where pseudolabels can be derived.
Sycophancy Hides Linearly in the Attention Heads
aclanthology.org/2026.eacl-lo...
Sycophancy is a concerning phenomenon which LLMs regularly exhibit. Showing where it is encoded, and also more surprisingly, showing that it is in the attention heads, is very cool.
To ease my FOMO from not attending @eaclmeeting.bsky.social, I skimmed the proceedings while playing the Tangier episode of Parts Unknown.
I'll do something different and shout out 5 (subjectively) interesting works of authors I'm *not* closely related to, in no specific order:🧵⬇️
Thinking of applying for an #MSCA Postdoctoral Fellowship in 2026?
I’m open to supervising at Bocconi! Feel free to reach out.
By submitting an expression of interest to Bocconi, selected applicants will receive full proposal support.
🗓️ Deadline: April 15
👉 www.unibocconi.it/en/horizon-e...
Excited to present this work together with @dippedrusk.com at #EACL. Join us in the poster session 1 (11:30-13:00) 🔥
Excited to share that @milanlp.bsky.social will be presenting 5 new papers at #EACL2026 and workshops in Rabat 🇲🇦!
Argh this sucks. apparently they lost their funding guarantee
You are #EACL2026? Check out some great work from the CSS Department @gesis.org, our data Science Methods team, with great collaborators!
I’m seeing close to zero reaction/conversation about this on here. This is huge news for open research on language models, especially in the US.
Yup this is a massive loss. OLMo (+entire ecosystem, OLMoTrace, the NeurIPS tutorial on the LM pipeline,...) was incredibly valuable and now I feel I took it for granted all this time. Not even counting all the great research coming out of AllenAI.
“It’s toddler AI misinformation at an industrial scale. It’s very risky for the developing brain.”
Children’s media experts say AI-generated “slop" has infiltrated the internet, preying on young children and their unsuspecting caregivers.
We'll have a reproducibility track at this years' Blackbox workshop! Details are still within a slightly opaque box.
We want to see if cleaning solutions that make opaque boxes 📦 transparent 🍱 work on different boxes 🎁📮🧰🥡, and with different 🧽solution-to-water🧼ratios!
Thank you Maria!
Check out our paper & code for the full results!
arxiv.org/abs/2603.03308
technion-cs-nlp.github.io/OldHabitsDie...
The probabilistic mechanism can also be calculated for closed models.
We find relatively similar probabilistic results compared to open models. Given the high correlation between our probabilistic and geometric results:
➡️ We could attempt to induce the geometry of closed models!
This correlation dissolves in an inconsistent conversations (spanning different topics).
This finding aligns with adversarial strategies that employ unrelated tokens to jailbreak models (Zou et al., 2023; Qi et al., 2025)
We bridge two worlds:
- Probabilistic: Modeling chats as Markov chains.
- Geometric: Measuring the orthogonality and dynamics in the internal state.
We find a high correlation between the two; the larger the probabilistic consistency - the larger the internal trap!
How does an LLM’s past influence its future?🤔
In new work, led by @adisimhi.bsky.social, together with @fbarez.bsky.social @boknilev.bsky.social and Shay Cohen, we find conversational history creates a latent "geometric trap" which makes old habits e.g. hallucinations hard to break!