Advertisement · 728 × 90

Posts by Dean W. Ball

https://www.hyperdimensional.co/p/where-we-are-headed

t.co/HoraJ0o77L

1 year ago 2 0 0 0
Post image

There are tons of technocratic AI policy proposals, but little concrete intuitions about the near-term future to motivate those proposals. The result is that the field sometimes feels disembodied, devoid of a concrete vision.

Here is what I think is coming.

1 year ago 9 2 1 1
Preview
Sharing the latest Model Spec We’ve made updates to the Model Spec based on external feedback and our continued research in shaping desired model behavior.

OpenAI releasing a model spec is one of my favorite things about them. All frontier labs should be doing this. Closes the gap between intentions and results, so you can interpret them better.

openai.com/index/sharin...

1 year ago 20 4 0 0
https://www.hyperdimensional.co/p/diffusion-confusion

Enjoy!

t.co/eQczeWf4Vl

1 year ago 4 0 0 0
Post image

My latest, on the AI export controls generally and the diffusion rule in particular.

This regulatory adventure, whatever you think of it on the merits, is a case study in how rules beget more rules—how regulation grows, like a plant or a weed.

1 year ago 11 1 1 0
Post image

Excited to launch a new podcast with @deanwb.bsky.social about AI. It's called AI summer, and can be found on all major podcast apps. The first episode is an interview with political scientist @jonaskonas.bsky.social about AI policy in the Trump era. www.aisummer.org/p/jon-askona...

1 year ago 19 5 0 0
Post image

A little late, but--my 2025 'lookahead' post is here, covering some of the major AI products, research, and policy I'll be watching this year. I hope you enjoy!

t.co/CZdR8zjsFH

1 year ago 8 0 0 0
Preview
The Coming Year of AI Regulation in the States | TechPolicy.Press American AI policy in 2025 will almost certainly be dominated, yet again, by state legislative proposals rather than federal government proposals.

American AI policy in 2025 will almost certainly be dominated, yet again, by state legislative proposals rather than federal government proposals, writes Dean Ball.

1 year ago 6 7 1 0
Advertisement

Hey, it’s worked well for my specific uses (which often involve broad surveys of state policies), but I have heard lots of people saying it is mid for them. Ymmv!

1 year ago 1 0 1 0

I think it mostly wouldn’t apply to the dod specifically but it absolutely would apply to any dod contractors with a presence in Texas.

1 year ago 3 0 0 0
Preview
Opinion | AI Can’t Teach AI New Tricks A lack of human prose leads to what’s called ‘model collapse’ and pure gibberish.

I would say I’ve heard these points made by people from many of the most prominent think tanks in dc. Would also point you to articles like this recent one from the Wall Street journal

www.wsj.com/opinion/ai-c...

1 year ago 3 0 0 0
Preview
One Down, Many To Go Reflections on year one of Hyperdimensional

Link here: www.hyperdimensional.co/p/one-down-m...

1 year ago 3 0 0 0
Post image

Hyperdimensional is almost a year old! So I wrote a reflection on the first year, covering the origins of the project, SB 1047, where I erred, the impossibility of controlling technocracy, where I see the project going next, and more. Enjoy!

1 year ago 5 0 3 0

ah man, we’re fucked

1 year ago 1 0 0 0

I really wish more people criticized my own work, too.

1 year ago 4 0 0 0

“Since we necessarily underestimate our creativity, it is desirable that we underestimate to a roughly similar extent the difficulties of the tasks we face, so as to be tricked… into undertaking tasks which we can, but otherwise would not dare, tackle.”

-Albert Hirschman

1 year ago 5 0 0 0
https://www.hyperdimensional.co/p/impact-assessments-are-the-wrong

Heres the post: www.hyperdimensional.co/p/impact-ass...

1 year ago 0 0 0 0

Everyone I know who follows AI policy closely knows about these bills and is concerned. But there is no public outcry, like there was with SB 1047.

It’s time to change that.

1 year ago 0 0 1 0
Advertisement

One is already law (going into effect in 2026) in CO. Another will probably be law soon enough, via bureaucrat “rule making,” in CA. It has a good chance of passing next year in TX, VA, and CT. We’ll see it elsewhere, too.

1 year ago 0 0 1 0

I’m referring to a slate of civil rights-based bills we will see throughout America. They’re among the most complex AI bills with a serious chance of becoming law.

1 year ago 0 0 1 0
Post image

America is charging into an AI policy regime that will slow down adoption of AI, shrink model developers’ addressable market, and could deter frontier AI investment.

It’s one of the worst ways to regulate AI, it could become a national standard soon, and it’s under-discussed.

1 year ago 1 0 1 0

Please share!

1 year ago 0 0 1 0
Post image
1 year ago 2 0 0 0

If chatbots stay in their current role—assistants I use for random tasks throughout the day—then ads have a negligible effect on their utility.

If however chatbots become something like what the ai companies all aspire to—genuine “virtual chiefs of staff”—ads may be ruinous.

1 year ago 2 0 2 0

The world we live in today is already incredibly Cyberpunk, by comparison with the world of the 90s and early 2000s that I grew up in

1 year ago 9 1 1 1

Evidently, yes. There is one item from the bulleted list that we omit in that passage (discrimination).

1 year ago 0 0 1 0

The post is just that: a post summarizing my own views. And my view is that the focus on nebulous risks in the CoP, in the AI Act itself, and in various American policy documents is a foundational error. Also, I’ll just note that you are wrong: we don’t say every listed risk is easily measurable.

1 year ago 1 0 1 0

I would suggest reading the articles you’re criticizing before you criticize them, in general. Neither miles nor i strawmanned anyone. if you think the current document is even close to well scoped, that’s fine! We just disagree.

1 year ago 0 0 1 0
Advertisement

I also do not think it’s right to say that the government (the people who can fine these companies billions) should “merely” tell developers to “consider” what amounts to a massive range of nebulous risks. There is nothing “mere” about that!

1 year ago 0 0 1 0

The portions we quoted and criticized are quoted directly from the piece. You can disagree with our criticism and believe these risks should be considered, but it is not a strawman to quote directly from a piece.

1 year ago 0 0 1 0