Advertisement · 728 × 90

Posts by Max Ilse

The result is a fair, end‑to‑end comparison that isolates what actually drives performance for radiology foundation models.

#AI #MedicalImaging #FoundationModels #ScalingLaws #Radiology

6 months ago 2 0 0 0

including not just findings but also lines & tubes classification/segmentation and report generation. We also test the effect of adding structured labels alongside reports during CLIP‑style pretraining, and study scaling laws under these controlled conditions.

6 months ago 0 0 1 0

That makes it hard to tell whether wins come from the model design or just from more data/compute or favorable benchmarks. We fix this by holding the pretraining dataset and compute constant and standardizing evaluation across tasks,

6 months ago 0 0 1 0

Why this matters: Prior comparisons of radiology encoders have often been apples‑to‑oranges: models trained on different datasets, with different compute budgets, and evaluated mostly on small datasets of finding‑only tasks.

6 months ago 0 0 1 0

✅ Pretrained on 3.5M CXRs to study scaling laws for radiology models
✅ Compared MedImageInsight (CLIP-based) vs RAD-DINO (DINOv2-based)
✅ Found that structured labels + text can significantly boost performance
✅ Showed that as little as 30k in-domain samples can outperform public foundation models

6 months ago 0 0 1 0
Preview
Data Scaling Laws for Radiology Foundation Models Foundation vision encoders such as CLIP and DINOv2, trained on web-scale data, exhibit strong transfer performance across tasks and datasets. However, medical imaging foundation models remain constrai...

🩻Excited to share our latest preprint: “Data Scaling Laws for Radiology Foundation Models”
Foundation vision encoders like CLIP and DINOv2 have transformed general computer vision, but what happens when we scale them for medical imaging?

📄 Read the full preprint here: arxiv.org/abs/2509.12818

6 months ago 5 2 1 0

What a damning abstract

11 months ago 5 0 0 0
Post image

I want to reshare @brandfonbrener.bsky.social's @NeurIPSConf 2024 paper on CoLoR-Filter: A simple yet powerful method for selecting high-quality data for language model pre-training!

With @hlzhang109.bsky.social @schwarzjn.bsky.social @shamkakade.bsky.social

1 year ago 18 8 2 1
Screenshot of 'SHADES: Towards a Multilingual Assessment of Stereotypes in Large Language Models.'
SHADES is in multiple grey colors (shades).

Screenshot of 'SHADES: Towards a Multilingual Assessment of Stereotypes in Large Language Models.' SHADES is in multiple grey colors (shades).

⚫⚪ It's coming...SHADES. ⚪⚫
The first ever resource of multilingual, multicultural, and multigeographical stereotypes, built to support nuanced LLM evaluation and bias mitigation. We have been working on this around the world for almost **4 years** and I am thrilled to share it with you all soon.

1 year ago 128 23 6 3
Advertisement
Preview
Econometrics and Data Science

We’re looking for a motivated researcher to apply for a Marie Skłodowska-Curie postdoc with our Econometrics & Data Science group at SDU!

Focus: Causal Inference, Machine Learning, Big Data
Full support for promising projects

More info & apply:
www.sdu.dk/en/om-sdu/in...

1 year ago 18 12 0 1
Preview
a cartoon pikachu says join us in white letters ALT: a cartoon pikachu says join us in white letters

Apply!
Assistant Professor (L/SL) in AI, including computer vision [DL 5 Mar] @BristolUni - awarded AI University of the Year in 2024.
DM to myself or @_SethBullock_ for inquiries (don't us send CV pls!, apply directly)
www.bristol.ac.uk/jobs/find/de...

1 year ago 20 9 0 0
Post image

🎓 💫 We are opening post-doc positions at the intersection of AI, data science, and medicine:
• Large Language Models for French medical texts
• Evaluating digital medical devices: statistics and causal inference

1 year ago 27 16 1 0
Preview
When A.I. Passes This Test, Look Out The creators of a new test called “Humanity’s Last Exam” argue we may soon lose the ability to create tests hard enough for A.I. models.

If you name your AI benchmark "Humanity's Last Exam" and get Kevin Roose to gush about it, you work in advertising, not in computer science.

www.nytimes.com/2025/01/23/t...

1 year ago 282 27 9 2

Happy to announce a collaboration with the Mayo Clinic to advance our research in radiology report generation!
newsnetwork.mayoclinic.org/discussion/m...
Tagging some of the core team: @valesalvatelli.bsky.social @fepegar.com @maxilse.bsky.social @sambondtaylor.bsky.social @anton-sc.bsky.social

1 year ago 7 2 1 0

The video of our talk "From Augustus to #Trump – Why #Disinformation Remains a Problem and What We Can Do About It Anyway" at #38c3, Europe's largest Hacker conference, was published, including the German original and English and Spanish translations:

media.ccc.de/v/38c3-von-a...

1 year ago 8 2 1 0

Internship in our group at Mila in reinforcement learning + graphs for reducing energy use in buildings.

More info and submit an application by Jan 13 here:
forms.gle/TCChXnvSAHqz...

Questions? Email donna.vakalis@mila.quebec with [intern!] in the subject line.

1 year ago 13 4 0 0
Screenshot of Table of Contents (Part 1)

Contents
1 Introduction 217
2 Positionality 221
3 Overview of Risks and Harms Associated with Computer
Vision Systems and Proposed Mitigation Strategies 223
3.1 Representational Harms . . . . . . . . . . . . . . . . . . . 223
3.2 Quality-of-Service and Allocative Harms . . . . . . . . . . 229
3.3 Interpersonal Harms . . . . . . . . . . . . . . . . . . . . . 237
3.4 Societal Harms: System Destabilization and Exacerbating
Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . 245
4 Frameworks and Principles for Computer Vision
Researchers 266
4.1 Guidelines for Responsible Data and Model Development . 267
4.2 Measurement Modeling . . . . . . . . . . . . . . . . . . . 271
4.3 Reflexivity . . . . . . . . . . . . . . . . . . . . . . . . . . 273
5 Reorientations of Computer Vision Research 276
5.1 Grounded in Historical Context and Considering
Power Dynamics . . . . . . . . . . . . . . . . . . . . . . . 276
5.2 Small, Task Specific . . . . . . . . . . . . . . . . . . . . . 279
5.3 Community-Rooted . . . . . . . . . . . . . . . . . . . . . 280

Screenshot of Table of Contents (Part 1) Contents 1 Introduction 217 2 Positionality 221 3 Overview of Risks and Harms Associated with Computer Vision Systems and Proposed Mitigation Strategies 223 3.1 Representational Harms . . . . . . . . . . . . . . . . . . . 223 3.2 Quality-of-Service and Allocative Harms . . . . . . . . . . 229 3.3 Interpersonal Harms . . . . . . . . . . . . . . . . . . . . . 237 3.4 Societal Harms: System Destabilization and Exacerbating Inequalities . . . . . . . . . . . . . . . . . . . . . . . . . . 245 4 Frameworks and Principles for Computer Vision Researchers 266 4.1 Guidelines for Responsible Data and Model Development . 267 4.2 Measurement Modeling . . . . . . . . . . . . . . . . . . . 271 4.3 Reflexivity . . . . . . . . . . . . . . . . . . . . . . . . . . 273 5 Reorientations of Computer Vision Research 276 5.1 Grounded in Historical Context and Considering Power Dynamics . . . . . . . . . . . . . . . . . . . . . . . 276 5.2 Small, Task Specific . . . . . . . . . . . . . . . . . . . . . 279 5.3 Community-Rooted . . . . . . . . . . . . . . . . . . . . . 280

Screenshot of Table of Contents (Part 2)

6 Systemic Change 285
6.1 Collective Action and Whistleblowing . . . . . . . . . . . . 285
6.2 Refusal/The Right not to Build Something . . . . . . . . . 287
6.3 Independent Funding Outside of Military and Multinational
Corporations . . . . . . . . . . . . . . . . . . . . . . . . . 289
7 Conclusion 291
References 293

Screenshot of Table of Contents (Part 2) 6 Systemic Change 285 6.1 Collective Action and Whistleblowing . . . . . . . . . . . . 285 6.2 Refusal/The Right not to Build Something . . . . . . . . . 287 6.3 Independent Funding Outside of Military and Multinational Corporations . . . . . . . . . . . . . . . . . . . . . . . . . 289 7 Conclusion 291 References 293

Dear computer vision researchers, students & practitioners🔇🔇🔇

Remi Denton & I have written what I consider to be a comprehensive paper on the harms of computer vision systems reported to date & how people have proposed addressing them, from different angles.

PDF: cdn.sanity.io/files/wc2kmx...

1 year ago 386 164 8 10
Posts, barbed wire fences and the main gate of the former Auschwitz II-Birkenau camp.

Posts, barbed wire fences and the main gate of the former Auschwitz II-Birkenau camp.

Help us commemorate victims, preserve memory & educate the world. Amplify our voice.

Your interaction here is more than just a click. It is an act of remembrance against forgetting. Like, share, or quote our posts.

Let people know that @auschwitzmemorial.bsky.social is present here.

1 year ago 1709 754 48 62
Advertisement
Video

New timeline, same problems, same solution

1 year ago 250 49 10 3
Post image

Kudos to @blackhc.bsky.social for calling out this neurips oral: openreview.net/forum?id=0NM... for not giving the rho-loss paper (arxiv.org/abs/2206.07137) the recognition it deserves!

1 year ago 8 0 1 1

In all of the reposts I see of articles criticising generative AI, I still don't see enough mention of work like Dr. Birhane's, which shows the biases against disadvantaged groups in the training datasets.
This is very good.

1 year ago 50 18 1 0
Preview
Probabilistic weather forecasting with machine learning - Nature GenCast, a probabilistic weather model using artificial intelligence for weather forecasting, has greater skill and speed than the top operational medium-range weather forecast in the world and provid...

The next generation of probabilistic machine learning for weather called GenCast is published in @natureportfolio.bsky.social today 🥳. Amazing to see the collective progress in ML for weather as a field over the last 5 years. 🏖️ www.nature.com/articles/s41...

1 year ago 157 29 3 6
Post image

One of the reasons the university sector has come so spectacularly off the rails is the fact it's so unfriendly to family life, people with caring responsibilities and parents. The attitude is often: 'Not working 24/7? You're not fully committed!'
www.science.org/content/arti...

1 year ago 94 12 8 2

This is just sad

1 year ago 3 1 0 0
Preview
Medical Adaptation of Large Language and Vision-Language Models: Are We Making Progress? Several recent works seek to develop foundation models specifically for medical applications, adapting general-purpose large language models (LLMs) and vision-language models (VLMs) via continued pret...

Medically adapted foundation models (think Med-*) turn out to be more hot air than hot stuff. Correcting for fatal flaws in evaluation, the current crop are no better on balance than generic foundation models, even on the very tasks for which benefits are claimed.
arxiv.org/abs/2411.04118

1 year ago 259 57 8 13

What a surprise (not!). Yet again ... poor evaluations of specialized medical LLMs result in overhyped claims relative to the base LLMs. #bioMLeval

1 year ago 76 14 1 1

Not sure about IRM thought. They early stopped their experiments to get the colored MNIST results.

1 year ago 1 0 1 0
Advertisement

Good repost would have completely missed this :)

1 year ago 1 0 0 0

Thanks for the pack! Can you please add me :)

1 year ago 1 0 1 0
Screenshot of the paper.

Screenshot of the paper.

Even as an interpretable ML researcher, I wasn't sure what to make of Mechanistic Interpretability, which seemed to come out of nowhere not too long ago.

But then I found the paper "Mechanistic?" by
@nsaphra.bsky.social and @sarah-nlp.bsky.social, which clarified things.

1 year ago 230 26 7 2