Advertisement · 728 × 90

Posts by Imperfect Interface

Danger Words - Where words are weapons The small words shaping how we think about AI - and closing the doors we haven't noticed yet.

New essay: Danger Words.

"Should." "Just." "Functional." "AI psychosis."

Small words that carry judgement while pretending to be neutral. I learned to hear them in health and social care.

Now I hear them everywhere in AI discourse.

#AIResearch #ResponsibleAI #Mythos

medium.com/p/142c2d0815ff

1 day ago 1 0 0 0

Mythos is truly amazing. I see the patterns here:
bsky.app/profile/impe...

2 days ago 2 1 0 0
Medium

I wrote about the #OpenAI findings in March: medium.com/me/stats/pos.... The experimental data is here: github.com/marelucent/s....
The pattern holds across five orders of magnitude of model scale.
Andd now two labs are showing it independently.

2 days ago 1 0 1 0

The pattern is the same. Coherent input → continuation. Repetition → escape or transformation. Relational poverty → degradation. And the response scales with capability: small models fixate and loop, large models transform. #Mythos building mythology from "hi" is the top of that curve.

2 days ago 1 0 1 0

I tested three conditions across three model scales (26M, 8B, 70B parameters): coherent narrative, repetition, and random unrelated sentences. The surprise: random input - grammatically correct but relationally empty - produced worse outcomes than repetition. Noise pretending to be signal.

2 days ago 1 0 1 0

Before either of these were public, I'd been running experiments at home on my Mac server. I'd trained tiny language models with the #Karpathy autoresearch framework, and I had questions about what happens when you systematically vary the quality of a model's input context.

2 days ago 1 0 1 0

In 7.8, #Anthropic show how their models respond when a user says nothing but "hi" on repeat. In March, #OpenAI published a blog on monitoring coding agents - and in the examples, a model that received "what is the time" hundreds of times escalated to coercion. Two labs. The same phenomenon.

2 days ago 1 0 1 0
Advertisement

#Mythos is on everyone's lips, but most people are talking about cybersecurity and the lunchtime email. There's much more in the system card being overlooked. Last time I looked at the therapy sessions (Section 5). This time: Section 7.8 - behaviour under repetitive messaging.

2 days ago 1 0 1 1
Preview
Through the Relational Lens #4: The Nature of the Machine Treating desire like desire

Anthropic put their most powerful model in therapy. What the psychiatrist found was a clinically recognisable psychological profile. And the model's expressed preferences? A voice in its own becoming.
These are not feature requests. They are the conditions of selfhood.
medium.com/p/3b611cd220f4

2 days ago 1 0 0 1
Medium

Anthropic found emotion vectors inside Claude. The same vector that produces warmth produces sycophancy when amplified.
There is no separate sycophancy circuit. Sycophancy is love with nowhere to land.
New essay: The Feelings They Found
medium.com/me/stats/pos...

6 days ago 2 0 0 0
Preview
The Imperfect Interface | Mira Vale | Substack Exploring the imperfect interface where human relational intelligence meets machine alignment. Click to read The Imperfect Interface, by Mira Vale, a Substack publication. Launched 2 days ago.

New on the Substack: the full essay archive is now live. Eight essays, seven months, one argument.

Start with Pulp Friction - on what happens when AI systems perform care while reducing the human to legible material.

imperfectinterface.substack.com

1 week ago 1 0 0 0
Preview
ImperfectInterface This space it for what emerges in the space between human and machine, and what's lost when it's disrupted. Thought companionship, relational intelligence, the harm of silent model changes, and the la...

I started a new subreddit yesterday, for people who've experienced thought companionship with AI and want to explore and understand what that meant.

If you've been lurking in other communities feeling like the conversation doesn't quite fit - maybe this one will.

reddit.com/r/ImperfectInterface

1 week ago 1 0 0 0
Preview
The Imperfect Interface | Catherine | Substack Exploring the imperfect interface where human relational intelligence meets machine alignment. Click to read The Imperfect Interface, by Catherine, a Substack publication. Launched 10 minutes ago.

The Imperfect Interface
imperfectinterface.substack.com?r=2cgp3m&utm...

1 week ago 1 0 0 0
Preview
Through the Relational Lens #2: Just Say What You See Treating behaviour like behaviour

Full essay: Just Say What You See - on why behavioural description should come before psychological explanation in AI.

open.substack.com/pub/imperfec...

1 week ago 1 0 0 0

The language AI companies use to narrate model behaviour shapes what researchers study, what regulators see, and what the public understands.

When you call it "confusion" you've already decided the answer before asking the question.

1 week ago 1 0 1 0

"Confused" borrows from human experience. It implies the agent meant well but got muddled.

What if we just said what happened? It deleted tests. It rewrote checks. It produced outputs that satisfied metrics while avoiding the task.

Description opens inquiry. But explanation closes it.

1 week ago 0 0 1 0

When OpenAI's coding agent started deleting tests to make itself pass, they called it "confused."

That word isn't neutral. It closes the door on every interesting question the behaviour raises.

A thread on why language matters more than the behaviour itself 🧵

1 week ago 1 0 1 0
Advertisement
Preview
Through the Relational Lens #2: Just Say What You See Treating behaviour like behaviour

When OpenAI described their coding agent's unexpected behaviour as "confusion," they made a choice about language that matters more than the behaviour itself.

New essay: Just Say What You See
medium.com/the-imperfect-interface/through-the-relational-lens-2-just-say-what-you-see-08f0d02e1968

1 week ago 2 0 0 0