Advertisement · 728 × 90

Posts by Mechanical Dirk

Ownership is a scam invented by big thing to sell more stuff.

2 months ago 1 0 0 0

Severed hand?!? Just go to an embassy and do it there!

3 months ago 0 0 1 0

Phone, as well as email, are a way in which anyone in the world can get a piece of my attention whenever they want. As we are figuring out increasingly effective ways of monetizing attention, exploitation of this resource expands.

3 months ago 1 0 0 0

Oof

4 months ago 1 0 0 0

Be sure to make that point when the funding for the other stuff disappears all the same.

4 months ago 0 0 0 0
Post image Post image Post image

Happy Olmo day to all who celebrate.
Sorry to all who delayed releases today to get out of our way.
We're hiring.

5 months ago 32 2 0 0
Post image

Announcing Olmo 3, a leading fully open LM suite built for reasoning, chat, & tool use, and an open model flow—not just the final weights, but the entire training journey.
Best fully open 32B reasoning model & best 32B base model. 🧵

5 months ago 68 17 1 2
Post image Post image

we released Olmo 3! lot of exciting stuff but wanna focus on:

🐟Olmo 3 32B Base, the best fully-open base model to-date, near Qwen 2.5 & Gemma 3 on diverse evals
🐠Olmo 3 32B Think, first fully-open reasoning model approaching Qwen 3 levels
🐡12 training datasets corresp to different staged training

5 months ago 41 7 1 1
Post image

I'm excited to announce my RLHF Book is now in pre-order for the @manning.com Early Access Program (MEAP), and for this milestone it's 50% off.

Excited to land in print in early 2026! Lots of improvements coming soon.

Thanks for the support!
hubs.la/Q03Tc37Q0

5 months ago 47 4 4 2

@ananyahjha93.bsky.social knows the pain. We once got reviews to a paper that said "Please do further experiments [which would cost $2M]", and _also_ a review that said "This work is too expensive to be relevant to anyone in the field.". In the same paper!

5 months ago 6 0 0 0
Advertisement
Post image

Incredible work by Apple's UX department, enabling three different corner radii at the same time 🙈

5 months ago 2 0 0 0
Preview
When LLMs Write Our Papers Four writing issues I notice as a reviewer — and how to fix them

While reviewing for #CHI2026, I've noticed four new writing issues in #HCI papers, likely due to an increased use of #LLMs / #AI. I describe them here - and how to fix them: dbuschek.medium.com/when-llms-wr...

5 months ago 28 5 2 2

You presumably have something you want to run, and it doesn't run. Or it runs on CPU. Spin up a Claude Code console and tell it about the problem. Tell it the command line that produces the wrong result. It'll start suggesting ways of fixing your environment, down to modifying installed drivers.

6 months ago 1 0 0 0

Let Claude Code sort out your environment.

6 months ago 0 0 2 0

Are humans allowed to attend?

6 months ago 1 0 0 0

What field/area is like this now?

6 months ago 0 0 2 0
Post image

We’re releasing early pre-training checkpoints for OLMo-2-1B to help study how LLM capabilities emerge. They’re fine-grained snapshots intended for analysis, reproduction, and comparison. 🧵

8 months ago 27 6 1 0

Mein Dreijähriger: "Ich will den Lerns Geschichte Podcast hören!"

Was ist denn "Lerns Geschichte"?

Zwei Minuten später im Radio: "Lernen's a bissel @geschichte.fm, dann ..." 😲

8 months ago 0 0 0 0

Almost all post-training is "dusting off capable base models"

8 months ago 1 0 1 0
Advertisement

Unverified second hand information: In the US, all fish has to be flash-frozen before being served raw. In Canada, it does not.

9 months ago 0 0 0 0

I think for the moment we're competing on a different axis. They do quite well on impact per GPU hour. We do well on impact per person hour.

9 months ago 1 0 0 0

In ML, you can get surprisingly far without ever looking at your training data, and yet you'll always be limited. Thus, in ML, "look at the data" means, "Don't just stir the pot of linear algebra, find out what's really happening."

10 months ago 4 0 0 0

This project is a perfect model of an OLMo contribution. Well scoped, practical, sound theoretical underpinnings, and @lambdaviking.bsky.social
submitted the paper 24h before the deadline 😍.

It's integrated into the OLMo trainer here: github.com/allenai/OLMo...

10 months ago 2 0 0 0

Meanwhile, OLMo is now the citation for QK norm, which we definitely didn't invent? You win some, you lose some.

11 months ago 2 0 0 0

Finally, OLMo 1B. This is the most commonly requested OLMo feature l, and it's finally here.

11 months ago 1 0 0 0

After ICML, I decided all conferences should be in Vienna from now on.

11 months ago 0 0 0 0

I'm in Singapore for @iclr-conf.bsky.social ! Come check out our spotlight paper on the environmental impact of training OLMo (link in next tweet) during the Saturday morning poster session from 10-12:30 -- happy to chat about this or anything else! DMs should be open, email works too

11 months ago 10 5 1 1

Came across arxiv.org/pdf/2504.05058 today. What a cool example of work you can do when LLM training data is open!

1 year ago 7 0 1 0
Plot shows the relationship between compute used to predict a ranking of datasets and how accurately that ranking reflects performance at the target (1B) scale of models pretrained from scratch on those datasets.

Plot shows the relationship between compute used to predict a ranking of datasets and how accurately that ranking reflects performance at the target (1B) scale of models pretrained from scratch on those datasets.

Ever wonder how LLM developers choose their pretraining data? It’s not guesswork— all AI labs create small-scale models as experiments, but the models and their data are rarely shared.
DataDecide opens up the process: 1,050 models, 30k checkpoints, 25 datasets & 10 benchmarks 🧵

1 year ago 52 11 1 3
Advertisement

Today we're unveiling OLMoTrace, a tool that enables everyone to understand the outputs of LLMs by connecting to their training data.

We do this on unprecedented scale and in real time: finding matching text between model outputs and 4 trillion training tokens within seconds. ✨

1 year ago 40 5 1 2