We know AI is biased, but what does that do to people's hiring decisions? (Spoiler alert: nothing good) To address this, @aylincaliskan.bsky.social and I wrote about legal implications and policy suggestions to protect people's rights and autonomy @brookings.edu
www.brookings.edu/articles/ais...
Posts by Kyra Wilson
And here is the presentation I gave on networking, self-promo, and how to make the most out of a conference. Hope this helps for everyone at NeurIPS!
www.youtube.com/watch?v=B9hG...
A picture of Katie Wilson on election night at her campaign party delivering a speech to supporters. she holds her hand over her heart
A picture of Katie Wilson at her election party on election night. speaking into a microphone to a crowd of supporters, there are blue balloons behind her on stage and gold and silver streamers
A black and white photo of Katie Wilson knocking on a door of a single family home while door knocking for her campaign
A photo of Katie Wilson looking at her phone while standing by the buzzer box for an apartment building while door knocking for her campaign
We took on a powerful incumbent who was expected to coast to reelection.
We faced more corporate PAC money than has ever been spent attacking a candidate in a Seattle election.
We built a people-powered movement rooted in hope for our city’s future.
And we won.
This is YOUR city!
Thanks for shout-out Samer, I’m glad our work resonated with you!
"No Thoughts Just AI: Biased LLM Hiring Recommendations Alter Human Decision Making and Limit Human Autonomy"
by @kyrawilson.bsky.social et al
It shows how people trust AI hiring proposals blindly, even when the AI's being racist. This highlights the importance to tackle algorithmic bias
Thanks to @uwnews.uw.edu for covering my + @aylincaliskan.bsky.social's recent work published at AIES 2025! www.washington.edu/news/2025/11...
🚨New paper: Reward Models (RMs) are used to align LLMs, but can they be steered toward user-specific value/style preferences?
With EVALUESTEER, we find even the best RMs we tested exhibit their own value/style biases, and are unable to align with a user >25% of the time. 🧵
If you’ve made it to the end of this thread, thanks for reading! I hope we can connect soon and chat about the role of all these works in making AI evaluation more valid, reliable, and actionable for the real-world! 🥳
I will present this tomorrow at 11:45am in Paper Session 6: Integrating AI into the Workplace. The preprint has more analyses of AI literacy's impact and discussion of the implications of this work for human autonomy in high-risk domains.
www.arxiv.org/abs/2509.04404
An image showing that people who completed an implicit association test before the resume-screening task were more likely to select Black and Hispanic candidates for high-status jobs and Asian candidates for low status jobs.
But a positive note is that when people took an implicit association test (commonly used for anti-bias training) before doing the resume-screening task, they increased their selection of stereotype-incongruent candidates by 12.7% regardless of how biased the AI model they interacted with was.
An image showing that people's decisions are very closely aligned with biased AI recommendations. Only in the most severe cases of AI bias do people make less biased decisions.
We showed people AI recommendations that had varying levels of racial bias and found that (at most) human oversight decreased bias in final outcomes by at most 15.2% which is still far from the outcome bias rates when no AI or unbiased AI was used.
3️⃣ Last but not least, I’ll be talking about a human-subjects experiment follow-up to my AIES 2024 paper with @aylincaliskan.bsky.social, also co-authored by Mattea Sim and Anna-Maria Gueorguieva!
arxiv.org/abs/2407.20371
I'll be discussion this work tomorrow at 3:15pm during Poster Session 4! Our preprint has additional findings and analyses comparing generated images to human baselines and measuring dimensions of skin tone other than lightness/darkness.
arxiv.org/abs/2508.17465
An image showing that Stable Diffusion v2.1 generates skin tones which are more diverse for a stigmatized racial identity versus Stable Diffusion XL which generates darker, more homogenous skin tones.
We also found that depictions of racial identities are getting more homogenized with successive releases of SD, reinforcing harmful ideas about what people with stigmatized identities "should" look like.
An image showing that Stable Diffusion XL has the darkest skin tones on average and stigmatized identities have darker skin tones that non-stigmatized identities.
We found that the newest model (SD XL) tends to generate images with darker skin tones compared to SD v1.5 and v2.1, but it still over-represents dark skin tones for stigmatized identities compared to non-stigmatized identities.
2️⃣ In another work with Sourojit Ghosh and @aylincaliskan.bsky.social, we analyzed skin tones in images of 93 stigmatized identities using three versions of Stable Diffusion (SD).
Chat with me about this work at Poster Session 2 (today at 6:15pm) or read our preprint in full!
arxiv.org/abs/2508.11067
An image showing the Fact Sheet we developed, which includes questions for the topics High-level Facts, Limitations and Ethical Considerations, Intended Use Cases, Implementation-level Details, and Maintenance and Lifecycle.
We also find that 89.4% of papers don’t provide detailed information about real-world implementation of their findings. Based on this, we made a Fact Sheet that to guide researchers in communicating findings in ways that enable model developers or downstream users to implement them appropriately.
1️⃣ Sourojit Ghosh and I conducted a review of AI bias literature and found that 82% do not provide an explicit definition of bias and 79.9% do not explore bias outside of binary gender bias. This means that only a particular subset of marginalized groups may benefit from AI bias research.
Happy to share that I’m presenting 3 research projects at AIES 2025 🎉
1️⃣Gender bias over-representation in AI bias research 👫
2️⃣Stable Diffusion's skin tone bias 🧑🏻🧑🏽🧑🏿
3️⃣Limitations of human oversight in AI hiring 👤🤖
Let's chat if you’re at AIES or read below/reach out for details!
#AIES25 #AcademicSky
Applying for a #PhD @ischool.uw.edu? Read 👇
Our student-run application feedback program will be open from October 20th through 1st November 2025.
Everyone applying, especially those from historically underrepresented groups or who have faced barriers in higher ed are highly encouraged to apply.
Tesla is being hit with $329m in damages for a crash in which the human driver says he knew Autopilot wasn't self-driving. This is super important, because it shows that Autopilot's design and marketing can induce inattention even when drivers consciously know they are supposed to pay attention.
📣 Exceptional resource alert from
@ucblaborcenter.bsky.social: a database of U.S. union bargaining provisions related to workplace tech. laborcenter.berkeley.edu/negotiating-...
Workers need a voice in how tech is utilized in the workplace— and here are 175 blueprints for what that looks like
There’s a lot of evil stuff. But these few sentences are a result of wild corruption, and would remove any guardrails against AI systems being used for massive civil and humans rights abuses on an unimaginable scale. Call your congresspeople, and scream about it in the streets.
I’m in this article along with other @uaw4121.bsky.social members to explain what’s happening to grad student researchers in the US right now.
It’s not looking great for us, and it’s looking even worse for America.
So: In 2017, Congress made this happen:
NIH: Proposed 22% cut --> 9% increase
NSF: Proposed 11% cut --> 4% increase
NOAA: Proposed 16% cut --> 4% increase
Obviously, 2025 is not 2017. A lot is diff now. But still:
💰 Congress, not WH, sets budgets.
📞 Public support & calls to Congress matter.
I wrote a post on how to connect with people (i.e., make friends) at CS conferences. These events can be intimidating so here's some suggestions on how to navigate them
I'm late for #ICLR2025 #NAACL2025, but in time for #AISTATS2025 #ICML2025! 1/3
kamathematics.wordpress.com/2025/05/01/t...
Excited to announce our #NAACL2025 Oral paper! 🎉✨
We carried out the largest systematic study so far to map the links between upstream choices, intrinsic bias, and downstream zero-shot performance across 131 CLIP Vision-language encoders, 26 datasets, and 55 architectures!