Advertisement · 728 × 90

Posts by Stephanie Hyland

Preview
Pancreatic cancer mRNA vaccine shows lasting results in an early trial Scientists caution that more research is needed, but nearly all of the patients who responded to the personalized vaccine are still alive six years later.

"Pancreatic cancer mRNA vaccine shows lasting results in an early trial: Scientists caution that more research is needed, but nearly all of the patients who responded to the personalized vaccine are still alive six years later."

4 days ago 9745 2908 156 602
The machines are fine. I'm worried about us. On AI agents, grunt work, and the part of science that isn't replaceable.

Hey, I wrote a thing about AI in astrophysics
ergosphere.blog/posts/the-ma...

3 weeks ago 1728 517 109 265

the greatest joy of being a computational scientist is having the computer work for you while you do something else

3 months ago 13 1 0 1

“Interpretability plays a special role in machine learning because instead of focusing on making the AI smarter, we focus on improving human insight. I think this is the most important category of interpretability research, and we do not do enough of it.”

😎😎😎

4 months ago 4 0 0 0
A poster titled “a circular argument” which has been cut into a circular shape

A poster titled “a circular argument” which has been cut into a circular shape

It’s a CIRCULAR poster! #eurips presenters innovating in poster design / fine motor skills

4 months ago 0 0 0 0
a hand-written poster on a poster board, featuring a hand-drawn QR code (the code does not work)

a hand-written poster on a poster board, featuring a hand-drawn QR code (the code does not work)

remember to always include a QR code on your poster. spotted at #eurips

4 months ago 5 0 1 0
Video

What coding with an LLM feels like sometimes.

4 months ago 267 64 10 6

when I ask candidates whether they've worked with "real medical data" this is the kind of thing that I mean

4 months ago 2 0 0 0

found a file from PhD days with the FORTY-EIGHT ways "ACE inhibitor" was encoded in the EHR system we were working wth

4 months ago 5 0 1 0

finally got around to booking my travel for #EurIPS2025! Looking forward to connecting with the European ML scene in Copenhagen

5 months ago 4 0 1 0
Advertisement

uv is so good

7 months ago 6 0 0 0
Post image

Some papers really have a good intro

7 months ago 16 1 4 0

The more rigorous peer review happens in conversations and reading groups after the paper is out with reputational costs for publishing bad work

8 months ago 48 5 2 3
Google's Gemini AI tells a Redditor it's 'cautiously optimistic' about fixing a coding bug, fails repeatedly, calls itself an embarrassment to 'all possible and impossible universes' before repeating 'I am a disgrace' 86 times in succession

Google's Gemini AI tells a Redditor it's 'cautiously optimistic' about fixing a coding bug, fails repeatedly, calls itself an embarrassment to 'all possible and impossible universes' before repeating 'I am a disgrace' 86 times in succession

I'll admit, I was skeptical when they said Gemini was just like a bunch of PhDs. But I gotta admit they nailed it.

8 months ago 7240 1653 70 158

what is the purpose of VQA datasets where text-only models do better than random?

8 months ago 1 0 0 0
Zotero screenshot showing four different papers with titles beginning with "MedAgent"

Zotero screenshot showing four different papers with titles beginning with "MedAgent"

lads can we stop

8 months ago 4 0 0 0
diagram from Anthropic paper with an icon & label that says “subtract evil vector”

diagram from Anthropic paper with an icon & label that says “subtract evil vector”

quick diagram of Bluesky’s architecture and why it’s nicer here

8 months ago 72 5 4 1

Emojis and massive try: except blocks. GitHub Copilot (at least Claude Sonnet 4) is very concerned about error handling.

8 months ago 2 0 1 0

if openreview were a lot fancier you could dynamically reallocate/cancel remaining reviews once a paper meets that expected minimum.

ideally you would mark these remaining reviews as optional rather than fully cancelled, in case that reviewer has already done work

8 months ago 3 0 0 1

it's frustrating how inefficient review assignments are: we target a minimum number of completed reviews per paper but in accounting for inevitable no-shows, some people end up doing technically unnecessary (if still beneficial) reviews

8 months ago 1 0 1 0
Advertisement

How many AI researchers fold their own laundry?

8 months ago 2 0 0 0
Post image

I am in the UK so feel free to discard, but I recently noticed Discord asking for age verification for some channels:

8 months ago 0 0 0 0
Preview
microsoft/maira-2-sae · Hugging Face We’re on a journey to advance and democratize artificial intelligence through open source and open science.

ALSO we have released the SAEs we trained, and the automated interp for all(!!)* features:
huggingface.co/microsoft/ma...

*all features for a subset of SAEs, we didn't run the full auto-interp pipeline on the widest SAE

9 months ago 4 0 0 0

We also found that the majority of the SAE features remained "uninterpretable", indicating room for improvement both in automated interpretability (we focused primarily on textual features!), but perhaps also questioning the SAE training and modelling assumptions. More work to be done here ✌️

9 months ago 2 0 1 0

... and in some cases we were able to steer MAIRA-2's generations, selectively introducing or removing concepts from its generated report.

But steering worked inconsistently! Sometimes it did nothing, or introduced off-target effects. We still don't fully understand when it will work.

9 months ago 1 0 1 0

We found interpretable and radiology-relevant concepts in MAIRA-2, like:
- "Aortic tortuosity or calcification"
- "Placement and position of PICC lines"
- "Presence of 'shortness of breath' in indication"
- "Describing findings without comparison to prior images"
- "Use of 'possible' or 'possibly'"

9 months ago 1 0 1 0
Post image

We performed the full pipeline of SAE training, automated interpretation with LLMs, steering, and automated steering evaluation.

9 months ago 1 0 1 0
Preview
Insights into a radiology-specialised multimodal large language model with sparse autoencoders Interpretability can improve the safety, transparency and trust of AI models, which is especially important in healthcare applications where decisions often carry significant consequences. Mechanistic...

New work from my team! arxiv.org/abs/2507.12950
Intersecting mechanistic interpretability and health AI 😎

We trained and interpreted sparse autoencoders on MAIRA-2, our radiology MLLM. We found a range of human-interpretable radiology reporting concepts, but also many uninterpretable SAE features.

9 months ago 11 4 1 0

Mexico is an *official* NeurIPS event, it’s an additional location for the conference and is different to the endorsement of EurIPS.

9 months ago 1 0 1 0
Advertisement

It’s an endorsed event but is not actually officially NeurIPS! Maybe if this experiment works well there will be more distributed (official) NeurIPS locations in future.

9 months ago 1 0 1 0