Specifically, I believe many things get worse when you try to optimize them because the underlying assumptions aren’t robust to the amount of computational power we are able to leverage.
Posts by Stella Biderman @ ICLR
My hot take is that the median social system breaks under too much optimization pressure, and we should stop trying to optimize things
I can’t even come up with a way to get to 600% here… the obvious error to make would be 100*(600/10) but that gives 6,000% not 600%.
Excited to be on my way to @iclr-conf.bsky.social! Come stop by our posters and hit me up. I'm especially excited to talk about
- Open weight safety
- Training dynamics and interpretability over time
- Memorization and machine unlearning
- Open data
- Rigorous experimental design
**702 sorry
FISA 207 is blatantly illegal and immoral and has always been obviously so. Republicans are pretending to not know this, just like Democrats did during the Biden administration.
This is bipartisan evil.
Feb 3, 2025 - We started fighting to save our data.
July 3, 2025 - We launched #SaveOurSigns with Minn librarians.
April 2026 - We are still talking about the importance of public data as a public good.
❤️🛟
1. No, it just doesn’t provide evidence for the claim that that’s happening
2. I’m much more worried about the dangers of right winger extremism than moderating the left. I think if the reported phenomenon was real it would still be a net positive for society
Regretfully, the story about LLMs anti-polarizing people was not real.
We can only hope 🙏
We currently live in a world that is a lot worse than it could be because of social media created echo chambers and algorithmic promotion of political extremism. So yes, stopping models from doing this is a huge win for safety.
If this is real, it’s very plausibly the biggest win for alignment research.
OLMoCR is kinda mediocre in my testing. What kinds of documents has it been good for you in?
They have a 175M / 4 year grant from the NSF and NVIDIA that’s earmarked for open source AI.
It’s really good
You have a moral imperative to refuse to work with these people or develop models for these purposes.
I’m having trouble figuring out what it would mean for someone to not let you FT your open weight model. Is this about pre-release evaluation?
That would be quite helpful!
I would be very interested in seeing a talk script and the resulting slides side by side. I tried taking your advice with Claude and was pretty disappointed in the resulting slides tbh
I’m not sure how to make an argument like this / I’m not 100% sure what you’re looking for.
How do you identify which problems are interesting and valuable? When people don’t work on problems that matter, why do you think that is?
Do you have a reference for how to do "a bound derived from a differential approach"?
If I was going to claim that a finetuning methodology for machine unlearning “really worked,” what evidence would you like to see?
I'm not sure if I'm more called out by this skeet or the fact that I've had two kidney stones already tbh...
Examples of mislabeled web text by existing LangID systems. A full text version is available on the blog post below.
Examples of mislabeled web text by existing LangID systems. A full text version is available on the blog post below.
Language identification still proves to be a challenging task, especially for web data. In collaboration with @mlcommons.org @eleutherai.bsky.social @jhu.edu and 97 community members, we created CommonLID, a new benchmark for LangID for 100+ languages!
Has anyone else had Claude code become non-functional recently? Even with a test input it spins for minutes without doing anything. Same thing happens in terminal.
It's going to get worse because people hate AI
The only reasons I use social media platforms is to get eyeballs on research and to yell at people who are wrong online.
X > Bluesky at both for me