Advertisement ยท 728 ร— 90

Posts by

๐Ÿšจ Deadline Extended to Feb 5 (AoE)!
CFP still OPEN for the #AFAA2026 Workshop at @iclr-conf.bsky.social โ€” on fairness across alignment & agentic AI systems.
Full & tiny papers welcome โ€ข Interdisciplinary work encouraged!
๐Ÿ”— afciworkshop.org

#ICLR2026 #AFAA2026

2 months ago 1 1 0 1
Preview
AFAA 2026 The Algorithmic Fairness Across Alignment Procedures and Agentic Systems (AFAA) workshop aims to spark discussions on rethinking fairness in AI alignment procedures and agentic system development.

๐Ÿšจ CFP OPEN! Weโ€™re launching the #AFAA2026 Workshop at @iclr-conf.bsky.social on ๐—ณ๐—ฎ๐—ถ๐—ฟ๐—ป๐—ฒ๐˜€๐˜€ ๐—ฎ๐—ฐ๐—ฟ๐—ผ๐˜€๐˜€ ๐—ฎ๐—น๐—ถ๐—ด๐—ป๐—บ๐—ฒ๐—ป๐˜ ๐—ฎ๐—ป๐—ฑ ๐—ฎ๐—ด๐—ฒ๐—ป๐˜๐—ถ๐—ฐ ๐—”๐—œ ๐˜€๐˜†๐˜€๐˜๐—ฒ๐—บ๐˜€.
Submit your latest ideas (full or tiny papers!)
Interdisciplinary work especially welcome :D
๐Ÿ—“ Deadline: Jan 31 (AoE) | ๐Ÿ”— www.afciworkshop.org

#AFAA2026 #ICLR2026

3 months ago 6 2 0 4

Four case studies with the gap between the reality of model use and their sandbox evaluations in audits... Definitely need to take a deeper dive, great presentation by Emily Black!

9 months ago 0 0 0 0

Evaluations in the way the model would be deployed vs evaluations in only controlled unrealistic settings!

9 months ago 0 0 1 0

Allowing companies to do isolated audits can lead to D-Hacking!! More robust testing is needed...

9 months ago 0 0 1 0

Legal frameworks tend to have control over allocative decisions (Yes/No outcomes), which fit well with traditional ML systems... But not with GenAI systems

9 months ago 0 0 1 0

Zollo et al: Towards Effective Discrimination Testing for Generative AI
#FAccT2025

9 months ago 1 0 1 0

Nuance of stereotype errors is so important to understand their true harms... Insightful presentation by @angelinawang.bsky.social

9 months ago 0 0 0 0

Women tend to report stereotype-reinforcing errors as more harmful while men tend to report stereotype-violating errors as more harmful...

9 months ago 0 0 1 0

Some items are more associated with men vs women (not surprising), but not all of them are equally harmful!!

9 months ago 0 0 1 0
Advertisement

Cognitive beliefs, attitudes and behaviours... Three ways to measure harms ('pragmatic harms')

9 months ago 0 0 1 0

Are all errors equally harmful? No! Stereotype-reinforcing errors vs stereotype-violating errors

9 months ago 0 0 1 0

Our understanding of stereotypes sometimes isn't indicative of reality.... they can appear in both directions, or might exist simply without harm

9 months ago 0 0 1 0

Wang et al: Measuring Machine Learning Harms from Stereotypes Requires Understanding Who Is Harmed by Which Errors in What Ways
#FAccT2025

9 months ago 1 0 1 0

Clear narrative and a great presentation by Cecilia Panigutti

9 months ago 0 0 0 0

Risk-measuring studies - Bringing it back to risk measurement, but this time with a clearly defined objective instead of risk-uncovering as before... Not just whether a risk exists, but 'how severe' is it?

9 months ago 0 0 1 0

Interface-design studies - Focus on UI design elements which impact user interaction

9 months ago 0 0 1 0
Advertisement

Reverse-engineering studies - Narrower scope and in-depth studies of how algorithms work... Methodological precision in the key!

9 months ago 0 0 1 0

Risk-uncovering studies - Typical starts from anecdotal evidence and help surface new risks

9 months ago 0 0 1 0

A review organized not by data collection technique, but by DSA risk management framework categories

9 months ago 0 0 1 0

Narrative review of algorithmic auditing studies, practical recommendation for best practices, and mapping to DSA obligations...

9 months ago 0 0 1 0

Panigutti et al: How to investigate algorithmic-driven risks in online platforms and search engines? A narrative review through the lens of the EU Digital Services Act
#FAccT2025

9 months ago 0 0 1 0

Such a broad topic... Excellent presentation by @feliciajing.bsky.social

9 months ago 0 0 0 0

Historical methods working alongside many other ways of auditing these models can help us take advantage of the broader scope of historical evaluations....

9 months ago 0 0 1 0

AI Audits have moved from bottom-up external evaluations to new age 'auditing companies'. While this has increased speed and scale, they have significantly narrowed the scope of auditing.

9 months ago 0 0 1 0

Why the history of AI assessments? A study through the lens of historical methods can help us understand neglected areas of auditing.

9 months ago 0 0 1 0

Sandoval and Jing: Historical Methods for AI Evaluations, Assessments, and Audits
#FAccT2025

9 months ago 0 0 1 0
Advertisement

Important recommendations on standardization of report creation and storage to allow better meta-analysis in the future... Eye opening presentation by @mkgerchick.bsky.social

9 months ago 0 0 0 0

Applicants impacted by these tools, whose demographic data is missing, are completely removed from these audits!

9 months ago 0 0 1 0

Serious issues with the data usage... most weird for me: 'simulated test data'!

9 months ago 0 0 1 0