Advertisement · 728 × 90

Posts by Nathan Lambert

This was one of those posts where you toil for a very long time around a point, in this case how open models are keeping up, in order to test every possible way your worldview could be wrong.

When done writing, it becomes obviously clear.

13 hours ago 3 0 0 0
Preview
Reading today's open-closed performance gap The complex factors that determine the single evaluation number so many focus on. Plus, how this changes in the future.

A TLDR is that unless the training dynamics of leading LLMs change or open model builders run out of money, this ~6 month performance gap from closed to open models is here to stay.
www.interconnects.ai/p/reading-to...

13 hours ago 17 3 1 1

A good thread about this from @juliekallini.bsky.social

x.com/JulieKallini...

4 days ago 5 1 1 2
My Bets on Where Open LLMs Go Next
My Bets on Where Open LLMs Go Next Open models are going through an incredible transition in relevance. This video is about reading the clues to what the long-term balance of open vs closed models will be. I've written 10+ articles…

New video! Talking through my 10+ open model pieces from early 2026 and how they fit together. They're all trying to figure out where open models go next.
www.youtube.com/watch?v=hKIX...

4 days ago 11 2 1 1
Post image

Opus 4.7 has a new tokenizer.
This means it's also a new base model.
Glory days of pretraining still very much going.

4 days ago 92 5 4 6
Post image

The current pace of token-efficient reasoning improvements across minor Claude Opus/GPT model versions is pretty wild. All signs point to this continuing.

4.6 to 4.7 could've been presented as a fairly large model bump in the past with this plot.

4 days ago 29 4 0 0

i've considered it, but I already maintain a private discord for my blog and I don't know if I have bandwidth for more socials.

5 days ago 2 0 0 0
Preview
My bets on open models, mid-2026 What I expect to come next and why, focused on the open-closed gap.

I spent some time trying to distill all the complex factors impacting open models -- economics, capabilities, distribution, policy, etc. -- into a clear list of beliefs. Here they are in full.
www.interconnects.ai/p/my-bets-on...

5 days ago 26 4 1 0

Do the work (and read my post contextualizing why all my recent efforts are important)!

6 days ago 1 0 0 0
Advertisement

It'll seem like a lot of your audience already knows everything if you've mentioned it once, but repetition is key. People are busy, algorithmic feeds are fickle, and people don't remember points/brands until they've seen it many times.

6 days ago 3 0 1 0

Many people, especially academics, too far too many projects and invest too little in each. The amount of work just on distributing each project is fairly high. If you don't work on distributing it, don't expect to get new readers.

6 days ago 1 0 2 0

It pays off so much on the downstream influence of your work to have direct audiences for it. These types of articles are a pure grind, but the directly convert hard earned reputation into success for projects.

6 days ago 1 0 1 0
Post image

One of my key strategies with Interconnects is to develop the practice of making my work obviously compelling to a wider audience, keeping them hooked over time and wondering what I'm up to, etc.
www.interconnects.ai/p/what-ive-b...

6 days ago 14 1 1 0

I'm going to add question & answer videos throughout the lecture to go deeper on topics that need it, and potentially cover some topics that are too recent and in flux to go in print. I expect 10-15 videos in total over the next few months.

6 days ago 4 0 0 0
Implementing RL Algorithms for LLMs | RLHF Course Lecture 4
Implementing RL Algorithms for LLMs | RLHF Course Lecture 4 YouTube video by Nathan Lambert

Implementing RL Algorithms for LLMs | RLHF Course Lecture 4
youtu.be/i-AIMpZHgeg

6 days ago 3 0 1 0
Understanding Policy Gradient Algorithms for RL on LLMs | RLHF Course Lecture 3
Understanding Policy Gradient Algorithms for RL on LLMs | RLHF Course Lecture 3 YouTube video by Nathan Lambert

Understanding Policy Gradient Algorithms for RL on LLMs | RLHF Course Lecture 3
youtu.be/K_Sj_-1BUMM

6 days ago 3 0 1 0
RLHF Foundations, IFT, Reward Modeling, Rejection Sampling | RLHF Course Lecture 2
RLHF Foundations, IFT, Reward Modeling, Rejection Sampling | RLHF Course Lecture 2 YouTube video by Nathan Lambert

RLHF Foundations, IFT, Reward Modeling, Rejection Sampling | RLHF Course Lecture 2
youtu.be/4gIwiSPmQkU

6 days ago 2 0 1 0
RLHF and Post-training Overview | RLHF Book Course, Lecture 1
RLHF and Post-training Overview | RLHF Book Course, Lecture 1 YouTube video by Nathan Lambert

RLHF and Post-training Overview | RLHF Book Course, Lecture 1
youtu.be/o6l6tJQgUg4

6 days ago 2 0 1 0
Welcome to The RLHF Book Course
Welcome to The RLHF Book Course YouTube video by Nathan Lambert

Playlist & welcome video: www.youtube.com/watch?v=jQPi...

6 days ago 3 0 1 0
Advertisement
Preview
Course | RLHF Book by Nathan Lambert Course lectures and talks on RLHF and post-training.

Excited to launch the accompanying free RLHF Course for my book. To kick it off, I've released:
- Welcome video
- Lecture 1: Overview of RLHF & Post-training
- Lecture 2: IFT, Reward Models, Rejection Sampling
- Lecture 3: RL Math
- Lecture 4: RL Implementation
Landing page: rlhfbook.com/course

6 days ago 76 21 6 1
Post image

a bit over 7 days out from the Gemma 4 release and it's models are outpacing (slightly) the equivalent Qwen 3.5 models on downloads. Big numbers!

1 week ago 17 0 0 0
Post image

Great stuff happening as we start to build out the codebases for my RLHF book (sorry, I haven't had much time until now!). Very accessible to issues, emails, comments etc to make it better. I'm also going to need another dgx spark.
github.com/natolambert/...

1 week ago 11 0 0 0
Preview
The inevitable need for an open model consortium And yes, I hate consortia too.

In 5-10 years, as models get more expensive & capable, I see the funding structures and support for open models breaking down. We need to consider if we need other options of supporting the open ecosystem.

The inevitable need for an open model consortium
www.interconnects.ai/p/the-inevit...

1 week ago 46 11 3 1
Explaining the Most Important Artemis II Photos
Explaining the Most Important Artemis II Photos GOOD STUFF Y'ALL All of the photos in this video were pulled from NASA Johnson's Flickr! https://www.flickr.com/photos/nasa2explore/ NASA's AMAZING video on Earth Rise:…

Incredible video. Very pure experience in why so many of us love science. www.youtube.com/watch?v=oaXR...

1 week ago 10 0 0 0

If you’re not concerned about ai progress in some way you’re not paying attention imo

1 week ago 2 0 0 0

A lot of it is that I don't want to mislead people in what is going to be a lasting book or not. Is a tradeoff of hype for imo a more grounded book.

1 week ago 2 0 0 0
Preview
Claude Mythos and misguided open-weight fearmongering Another dance around fears of open-source.

1. dont fall for anti open model fearmongering, but
2. acknowledge that AI capabilities are proceeding fast, and eventually there may be a reason to be more careful with open weight models

I don't think Mythos is that trigger, but I'm not 100% confident
www.interconnects.ai/p/claude-myt...

1 week ago 46 2 1 2
Preview
RLHF Book by Nathan Lambert The Reinforcement Learning from Human Feedback Book

Full book online: rlhfbook.com
Buy on Amazon: amzn.to/4cwCDJQ
Buy on Manning (currently big pre order discount): www.manning.com/books/the-rl...

1 week ago 8 1 0 0
Advertisement

Thank you for all your support. More projects related to the book being announced soon 🎥 I'm excited to reconnect with the community through in-person book events this summer and fall.

1 week ago 2 0 1 0

Any other topic is too dynamic, where writing a book today would be immediately outdated. RLHF is largely being overshadowed by lots of other developments in AI, but will always be around and at the forefront of human-AI interactions. The topic deserves coverage in depth and this platform.

1 week ago 2 0 1 0