Our ICML 2025 workshop on Actionable Interpretability drew massive interest. But the same questions kept coming up: What does "actionable" mean? Is it achievable? How?
We're ready to answer.
🧵
Posts by Sarah Wiegreffe
🥹🥰
Come join TRAILS as a postdoc at UMD (and work w folks at GW, MSU & Cornell) to conduct research and scholarship focused on approaches to AI that advance trust and trustworthiness with a great group of colleagues!
🌐 go.umd.edu/trails-postd...
🗓️ Summer/Fall 2026 start
If you're at #ICML2025, chat with me, @sarah-nlp.bsky.social, Atticus, and others at our poster 11am - 1:30pm at East #1205! We're establishing a 𝗠echanistic 𝗜nterpretability 𝗕enchmark.
We're planning to keep this a living benchmark; come by and share your ideas/hot takes!
I am also recruiting PhD students @univofmaryland.bsky.social for fall 2026 with interests in (causal/mechanistic) LM interpretability and its practical applications (steering, efficient adaptation, model editing, textual explanations for users, etc.).
I am at #ICML2025! 🇨🇦🏞️
Catch me:
1️⃣ Presenting this paper👇 tomorrow 11am-1:30pm at East #1205
2️⃣ At the Actionable Interpretability @actinterp.bsky.social workshop on Saturday in East Ballroom A (I’m an organizer!)
This week is #ICML in Vancouver, and a number of our researchers are participating. Here's the full list of Ai2's conference engagements—we look forward to connecting with fellow attendees. 👋
Thank you! Look forward to being colleagues.
Thank you!
Thank you!
Thanks :))
Thanks so much for all your support ☺️🥰
Thank you!
Thank you 😄
☺️ come visit!
A bit late to announce, but I’m excited to share that I'll be starting as an assistant professor at UMD CS @univofmaryland.bsky.social this August.
I'll be recruiting PhD students this upcoming cycle for fall 2026. (And if you're a UMD grad student, sign up for my fall seminar!)
Congrats Kristina! 😍
An image with the Vancouver skyline and the words "sign up to review". At the top are the logos of both the Actionable Interpretability workshop (a magnifying glass) and the ICML conference (a brain).
🚨 We're looking for more reviewers for the workshop!
📆 Review period: May 24-June 7
If you're passionate about making interpretability useful and want to help shape the conversation, we'd love your input.
💡🔍 Self-nominate here:
docs.google.com/forms/d/e/1F...
🤖: "Great review, but it could be improved by doing [exact thing I wrote in subsequent sentences]"
Where is version control and shared editing for keynote files?! 🤦♀️
We are quite excited about the leaderboard and release, and are open to feedback to help this remain a living benchmark.
Checkout our new preprint/project which has been over a year in the making! This has been a very fun collaboration (and one of the biggest I've personally participated in).
@amuuueller.bsky.social @boknilev.bsky.social and other co-authors are around #ICLR2025 if you want to find out more. 😊
See Yanai's thread for more info:
bsky.app/profile/yana...
2) On the connection between linear relational embeddings in LMs and frequency of relations in pretraining data
- Led by @jackmerullo.bsky.social w/ @nlpnoah.bsky.social @yanai.bsky.social
- arxiv.org/abs/2504.12459
- Yanai is presenting the poster tomorrow 04/26 10am-12:30pm (Hall 3+Hall 2B #236)!
I'm not at #ICLR2025, but have 2 works being presented:
1) Understanding how LMs answer multiple-choice questions
- arxiv.org/abs/2407.15018
- @boknilev.bsky.social is presenting the poster *now* until 12:30 (Hall 3+Hall 2B #207)
- & w/ @oyvind-t.bsky.social @hanna-nlp.bsky.social Ashish Sabharwal
I'm in Singapore for ICLR to present this paper:
Tomorrow, April 26th, 10-12:30 in Hall 3+2B #236
Come check it out!
arxiv.org/abs/2504.12459
💡 New ICLR paper! 💡
"On Linear Representations and Pretraining Data Frequency in Language Models":
We provide an explanation for when & why linear representations form in large (or small) language models.
Led by @jackmerullo.bsky.social, w/ @nlpnoah.bsky.social & @sarah-nlp.bsky.social
Have work on the actionable impact of interpretability findings? Consider submitting to our Actionable Interpretability workshop at ICML! See below for more info.
Website: actionable-interpretability.github.io
Deadline: May 9
📢 Open PhD Position in Interpretable Natural Language Processing at the Department of Computer Science, UCPH!
🗓 Application deadline is 15 January 2025.
Find more information about the position and apply here 👉 di.ku.dk/english/abou...
@apepa.bsky.social @iaugenstein.bsky.social