Advertisement · 728 × 90

Posts by Orpheus Lummis

thanks @teuber.bsky.social !

1 week ago 1 0 0 0
April 2026 - Montréal AI safety, ethics, governance Six events including Greywall agent sandboxing and AI persuasion research. Parliament hears from Bengio, Geist, and AIGS Canada on AI regulation. Tumbler Ridge lawsuit filed against OpenAI. Four new p...

Montréal AI safety, ethics, and governance newsletter, April 2026

- INDU committee launches AI regulation study
- Mila researchers find LLM agents can infer CoT monitoring
- 20/25 AI researchers flag automating AI R&D as top risk
- Multiple events!

aisafetymontreal.org/newsletter/2...

2 weeks ago 5 1 0 3
Preview
IVADO Thematic Semester - Statistical Foundations of AI Join IVADO Thematic Semester - Statistical Foundations of AI, May 4 to August 21, 2026. Learn more on Fourwaves.

IVADO (@ivado.bsky.social) offers two AI-safety-relevant upcoming workshops:

- Statistics in Trustworthy AI, May 11-15
- Uncertainty in AI, June 8-11

event.fourwaves.com/thematicseme...

3 weeks ago 0 0 0 0
Preview
The importance of AI character Forethought argues that AI character—e.g. how obedient, honest, or altruistic AI systems are—will shape power, conflict, and society far more than is recognized. Work to shape AI character could be hu...

New post: William MacAskill and Tom Davidson argue that AI character is a big deal.

Read it here: www.forethought.org/research/the...

3 weeks ago 1 1 0 0

Congrats!
Are there links the projects / repositories?

3 weeks ago 0 0 0 0
Preview
The Protopian Prize | The Fiction Contest The Protopian Prize is a fiction contest inviting you to share your vision of people working toward liberatory futures, meeting obstacles, and making real change. “Protopian”—a word coined by Kevin ...

Announcing The Protopian Prize | Fiction Contest 🕊️

Write the story of humanity’s future...

The Protopian Prize is a fiction contest inviting you to share your vision of people working toward liberatory futures, meeting obstacles, & making real change.

protopianprize.com

3 weeks ago 23 10 1 2
Preview
Public finance in the age of AI: A primer | Brookings In a new working paper, Anton Korinek and Lee Lockwood examine optimal taxation frameworks in the age of AI

UBI requires new tax revenue.

Korinek & Lockwood recently: As automation erodes labor income, consumption taxes take over. When AIs start reinvesting in their own expansion, tax that accumulation directly, balancing the rate between value for humans and growth.

www.brookings.edu/articles/pub...

3 weeks ago 6 0 1 0
Post image

Announcing the technical AI Governance Research (TAIGR) ICML workshop in July! Submissions (up to 8 pages) are due April 24. Co-submission with ICML and NeurIPS is encouraged.

taigr-workshop.com

3 weeks ago 2 2 0 0
Post image

This must-see new documentary is arriving in theatres this week. Through an honest and personal lens, Daniel Roher successfully highlights how each of us can move from passive observation to active contribution towards a more positive future with AI. www.youtube.com/watch?v=xkPb...

3 weeks ago 11 5 0 1
Advertisement

I‘m excited to present my work on Provably Safe Neural Network Control at @horizonomega.org‘s Guaranteed Safe AI online seminar on April 9th.
The talk will based on my NeurIPS‘24 paper with some updates on what I’ve been up to since :)
Feel free to join if you’re interested:

4 weeks ago 4 2 1 0
Preview
Provably Safe Neural Network Controllers via Differential Dynamic Logic · Zoom · Luma Provably Safe Neural Network Controllers via Differential Dynamic Logic Samuel Teuber – PhD Candidate, Institute of Information Security and Dependability…

Guaranteed Safe AI Seminars, April 2026:

Provably Safe Neural Network Controllers via Differential Dynamic Logic
Samuel Teuber – PhD Candidate, Institute of Information Security and Dependability (KASTEL), Karlsruhe Institute of Technology

Thursday, April 9, 1 PM EDT
RSVP: luma.com/920d2h7p

4 weeks ago 3 1 2 1
Post image

We are in Montréal, demanding frontier labs CEOs to commit to pausing AI frontier development, if the other labs do the same.

Nous sommes à Montréal, demandant que les PDGs d'IA s’engagent à suspendre le développement de l’IA frontière si les autres compagnies le font aussi.

4 weeks ago 4 2 0 0
Post image

B R O A D T I M E L I N E S
We should have neither short AI timelines, nor long timelines, but a broad probability distribution over when transformative AI will arrive.
My new essay explains why & explores the implications of such deep uncertainty.
🧵 1/

1 month ago 19 2 1 2
Preview
AI Control Hackathon · Luma This is the Montréal edition of the AI Control Hackathon. Schedule: Friday eveningIntro and dinner Saturday at FoulabHackathon day Henri Lemoine (Mila,…

AI Control Hackathon this weekend!

Given a misaligned model that may be actively trying to subvert safety measures, how can we design protocols that prevent catastrophic outcomes?

RSVP: luma.com/mhitd3xv

1 month ago 1 0 0 0
Preview
Manifestation de PauseAI devant Google pour arrêter la course à l’IA | PauseAI Demonstration at Google to Stop the AI Race · Luma FRANÇAIS: Nous allons manifester aux bureaux de Google à Montréal pour revendiquer que leurs PDGs, Sundar Pichai et Demis Hassabis, s’engagent publiquement à…

Joignez-nous à Montréal ce samedi 13-15h aux bureaux de Google, pour demander aux PDGs d'arrêter la Course à l'IA!

Join us in Montréal this Saturday 1-3pm at Google's offices, to demand the CEOs to Stop the AI Race!

luma.com/vw3nk8e6?tk=...

1 month ago 6 1 0 0
Preview
Solar Storms — LessWrong Most of civilization's electricity is generated far off-site from where it's delivered. This is because you don't want to be running and refueling co…

New post! Solar storms are damaging and expensive, are a tail risk for catastrophic harm, and can be averted straightforwardly and cheaply (only we haven't done so).

www.lesswrong.com/posts/ghq9Ew...

1 month ago 49 7 2 3
March 2026 - Montréal AI safety, ethics, governance AI Control hackathon with Apart and Redwood, Mila youth safety hackathon, and five events in Montréal. Bengio chairs the second International AI Safety Report. Six new papers from Mila, McGill, and Ud...

Montréal AI safety, ethics, and governance newsletter, March 2026 edition

- Intl. AI Safety Report: risk mgmt still voluntary
- 5 Montréal AI safety events this month
- CIFAR puts $1M toward alignment research
- Local papers on interpretability & hallucinations

aisafetymontreal.org/newsletter/2...

1 month ago 7 3 0 0

Within the next year we will have superforecaster-level AI. Their predictions would spread in the news, policy, planning, markets. But LLMs are highly correlated, so their shared biases and correlated failures like systematic overconfidence would propagate further into our collective epistemics.

1 month ago 1 0 0 0
Advertisement
Preview
Anyone Else Have Those Weird Dreams Where Sobbing Future Generations Beg You To Change Course? The human subconscious is such an interesting thing. No matter how much you think you’ve got it figured out, it’ll always spit out the most random stuff. Take me, for example. After coming home from a...

Commentary: Anyone Else Have Those Weird Dreams Where Sobbing Future Generations Beg You To Change Course?

1 month ago 7704 1429 89 63

Ran the Qwen 3.5 MoE family (3B–17B active params) on 155 recent prediction questions from ForecastBench. All are not well calibrated: overconfident when predicting near 100%, and many predictions clustered around 50% (hedging/low sharpness).

1 month ago 0 0 1 0

In 2006, DARPA had a research program (HI-MEMS) on implanting electrodes into insects during metamorphosis, so developing tissue would integrate them, to control their locomotion remotely.

1 month ago 6 0 0 0

Another approach which may be cleaner is using t-of-n threshold cryptography, where the PDS is one of the n shareholders but can never meet the threshold alone. Whenever a user wants to write to the PDS, their device co-signs.

FROST does this and is a standard as of 2024 in RFC 9591.

1 month ago 1 0 1 0

An active user might do hundreds+ signed commits to a PDS in a session (post, reply, liking, following, etc).

Self-hosting a PDS is inconvenient and unreliable relative to using specialized hosting services.

A path forward may be *short-lived delegated signing keys*, with user owning root keys.

1 month ago 2 0 1 0

The AI public benefit corporations do have humanity as their stated duty. Unfortunately, what they actually target is "what is tolerable by American law".

All the other AI companies are traditional corporations, which structurally do not even target the public benefit.

1 month ago 0 0 0 0

There was UN Secretary-General's High-level Advisory Body on Artificial Intelligence, established in 2023 with members from 33 countries, which released its final report "Governing AI for Humanity" in September 2024.

Its first recommendation was the creation of this Scientific Panel.

1 month ago 0 1 1 0
Advertisement

We need international red lines to prevent unacceptable AI risks.

Ban AI towards lethal autonomous weapons, mass surveillance, nuclear command & control, bioweapon assistance, unsupervised control of critical infrastructure, disinformation, CSAM, social scoring, and recursive self-improvement R&D.

1 month ago 4 0 0 0

Out of curiosity I asked Claude Opus about contemporary techniques vs this problem space. It created this web app comparing different methods claude.ai/public/artif... which you may find interesting

1 month ago 1 0 0 0

early physics of the mind fire

1 month ago 0 0 0 0