2/ With Tianmin Shu, Seth Lazar and Dylan Hadfield-Menell: we're building institutions that let real communities, not model developers, define how AI behaves.
Posts by Gillian Hadfield
1/ Who decides how AI systems behave? With Tianmin Shu, @sethlazar.org, and @dhadfieldmenell.bsky.social, we were named runner-up in the LaudeInstitute's inaugural Moonshot program and awarded a seed grant to find out. www.laude.org/moonshots
4/ This is a first step, not the finish line. We don’t have adoption of IVOs yet. But a state legislature voting unanimously for a new model of AI governance, and funding the evaluation, tells you something about where this is heading.
3/ Virginia is now studying whether to build it. I developed this framework in my book Rules for a Flat World and a 2019 paper applying it to AI safety. Watching a legislature take it seriously, fund the evaluation, and vote unanimously is not something I take for granted.
2/ The idea: independent bodies, licensed by the state, that would verify AI systems satisfy safety criteria set by the state. Voluntary for companies. Verifiers answer to the state, not to industry. Building on methods we've used in financial auditing and product safety.
1/ Something I’ve been working toward for a long time. Virginia just signed the first state legislation directing a formal study of Independent Verification Organizations for AI. Bipartisan votes of 84-14 in the House, unanimous 40-0 in the Senate.
5/ Full paper, open access: jair.org/index.php/ja...
4/ Policymakers read social media to gauge opinion. AI models train on internet data. Both get a distorted picture. We need to change who gets heard, not what people think.
3/ Ideological media orgs amplify this by signaling the other side is more extreme than it really is. Recommender systems, optimizing for engagement, sort people into communities where the loudest voices dominate.
2/ When rhetoric heats up, moderates face a squeeze from both sides. Allies speaking loudly substitute for your voice. Opponents shrink the reward from speaking up. Either way, moderates go quiet. No opinion change required.
1/ Social media makes it look like the public is deeply divided. But what if most of the public just isn’t speaking? In a new paper with Atrisha Sarkar in JAIR, we show why. We call it rational silence.
Democracy isn't a rulebook. It runs on daily interactions where people comply with norms and hold each other accountable. AI agents are about to join that system. We need to build them to read it. New paper with Rakshit Trivedi and Dylan Hadfield-Menell.
At the @FAR.AI London Alignment Workshop I made the case for Independent Verification Organizations: licensed, competing private entities that can grow our regulatory capacity to match the pace of AI capabilities. Full talk: bsky.app/profile/far....
Governments can’t translate “fair” or “safe” into technical specs fast enough. But leaving details to industry means the public loses its say. Regulatory markets close both gaps: governments set outcomes, private regulators compete to achieve them.
AI systems are quickly becoming embedded throughout the economy. But we have almost none of the regulatory tools, regulatory markets among them, to manage them. Here's what I think we should do about it: www.americanbar.org/groups/scien...
“The most practical governance framework currently in circulation.” That’s Forbes on the Independent Verification Organization model Fathom and I have been developing. Legislation takes years; IVOs move at the pace of innovation.
"Why not work on what kind of new governance is needed to ensure secure, reliable, predictable use of all frontier models, from all companies?"
In London today and tomorrow for the Alignment Workshop organized by FAR.AI. Keynoting alongside Rohin Shah and Allan Dafoe. I look forward to seeing everyone in attendance! www.far.ai/events/event...
The 2026 AI Safety Report's biggest finding isn't the risks it catalogs. It's the evidence gap. We're trying to build AI governance with almost no science underneath. Massive investment in the research regulatory systems depend on is overdue. internationalaisafetyreport.org
Screenshot of a LinkedIn post by Jack Shanahan (Retired USAF; Project Maven/DoD JAIC; NCSI MIS; SCSP Defense Partnership), posted 3 hours ago. In the post, Shanahan weighs in on the Anthropic-Pentagon dispute, noting that despite his Project Maven background, he's sympathetic to Anthropic's position. He argues no current LLM should be used in fully lethal autonomous weapons systems, calling that a reasonable red line, and opposes mass surveillance of US citizens as a second red line. He criticizes the public nature of the dispute, calls the supply chain risk designation "laughable," questions invoking the DPA against the company's will, and advocates for shared government-industry-academia governance of frontier AI models.
Why not work on new governance...
NIST just launched an AI Agent Standards Initiative for identity, security, and interoperability. AI agents are becoming economic actors with zero legal infrastructure in place. We require businesses to register to operate. Why expect less of AI agents? buff.ly/kTU2cfX
In Paris this week for IASEAI (Feb 24-26). Tuesday: panel on the International AI Safety Report. Thursday: keynote on regulatory markets, a panel on AI assurance, and a talk in Seth Lazar’s workshop on normative competence. If you’re at IASEAI, come say hello!
Congratulations to Yoshua Bengio and the 39 other experts appointed to the UN’s first Independent International Scientific Panel on AI. 117-2 in the General Assembly.
Better technology doesn’t fix broken institutions. The paper discusses regulatory markets as one path forward: instead of regulating providers directly, create a market for regulation itself. Worth a careful read. buff.ly/kbfvYqN
New in Lawfare from Justin Curl, Sayash Kapoor, & Arvind Narayanan: AI won’t automatically make legal services cheaper. I’ve been working on this for a long time, legal markets are broken because of adversarial dynamics, credence goods problems, & regulations that protect incumbents, not consumers.
Billions going into building AI, barely any into making sure it works for us. Talked with @kevintfrazier.bsky.social & Andrew Freedman about our proposal making its way through state legislatures to build a competitive market for AI oversight. New @scalinglaws.bsky.social podcast:
6/ Led by Shuhui Zhu with Yue Lin, Shriya Kaistha, Wenhao Li, Baoxiang Wang, Hongyuan Zha, and Pascal Poupart across Waterloo, Vector Institute, CUHK-Shenzhen, and Tongji. arxiv.org/abs/2602.07777
5/ We don't need AI agents that default to "nice." We need agents that understand when cooperation makes sense and when it doesn't. That takes institutional structure, not just training. Gossip turns out to be surprisingly powerful institutional structure.