congrats!!!
Posts by jessica dai
so close! that's standard error ❤️
i'm still pissed about this like the difference is literally too small to have been distinguishable with swe bench (500 samples) lmaoooo
hey wasn't this the same company that made a beautiful shiny "research" post about how AI evals should include error bars or something like that. or did they decide the CLT didn't apply here
I will be at ICML in a few weeks & would love to chat about how to make this real - I am a critic at heart and also hate self-promo so that’s how you know I really believe in this 🥲
various ways to read more 😀
blog post- argmin.net/p/individual...
position paper- arxiv.org/abs/2506.18133
fairness-oriented instantiation- arxiv.org/abs/2502.08166
& many thanks to brilliant collaborators
@rajiinio.bsky.social @irenetrampoline.bsky.social @beenwrekt.bsky.social & paula gradu !!
lots of other stuff I won’t get into rn (e.g., I think this is a prereq to any serious attempt at “democratic” AI!), and there’s also a ton of open research questions (stats, econ/ml, empirical methods, hci, …)
the core concept is individual reporting as a means to build collective knowledge. if one person has a bad experience, that doesn’t necessarily mean that there’s something wrong with the system — but if lots of people start reporting similar things, maybe we should pay attention.
we’ve already seen this informally with the chatgpt sycophancy debacle — a few days of twitter virality resulted in action and statements from openai — but what other, subtler, patterns are happening? what could we discover if we had better ways to listen to the public?
individual reporting for post-deployment evals — a little manifesto (& new preprints!)
tldr: end users have unique insights about how deployed systems are failing; we should figure out how to translate their experiences into formal evaluations of those systems.
right but one would hope that the date of doom _does_ get further away as safety research improves
bsky.app/profile/jess...
help ..
where are the bullshit "x% of experts believe" polls when you need them lol
well probably, but i wanna know how folks who do believe in that happening think about the field
or is it a secret third thing idk. scared to ask this on Real Twitter but genuinely curious how people think about the role of this field
like is it that the field has been ineffective (studied the wrong problems, advocated for the wrong positions, etc) or is it that every step of safety progress has been matched by 2 steps of capabilities progress (in which case, what are the best examples of safety work concretely reducing harm?)
perhaps this is a stupid question but given that ai safety has been a pretty vibrant (+ well funded) field for the last 5-10 years... how should we be thinking about the concern that (ai) catastrophe still is, allegedly, imminent
in middle school we were asked to write a short story in the style of edgar allan poe. as you might expect, all of our little pieces (even, especially, the ones the students thought were "good") were hilariously bad. anyway, i had forgotten about that homework until now
x.com/sama/status/...
back on bluesky to be mean about ai discourse
im ngl i think this kinda just means u are stupid
i don't work well under deadline pressure but i also don't work well without it. therefore,
... didn't we just talk about this ...
ill read it
The plan: Post your dissertation abstract online to rekindle a decades-long controversy about the utility of the humanities, turning your paper into the most-read publication in the history of your field
were you born yesterday
wait is that your house lmaooo
where
i just know these people would have been the biggest fans of japanese internment