Advertisement · 728 × 90

Posts by Julius Adebayo

Looks like Tesla’s models sometimes confuse train tracks with road lanes.

1 year ago 1 0 0 0
Post image

OLMo 2 tech report is out!

We get in the weeds with this one, with 50+ pages on 4 crucial components of LLM development pipeline:

1 year ago 47 9 3 0

The LCMs are cool though; however, it is early days. They give us a knob (concept representations) to understand and change the model's outputs. There is no reason why an LCM should also not have a COT (or be able to reason via search/planning)...we just have to ask it :)

1 year ago 1 0 0 0

The reasoning models are cool though; they explicitly enforce dependence on the model's cot, so here it should be a reliable explanation (? not sure tho). Played with 'thinking' gemini: it generate pages of COT sometimes, and now we have to figure what (and which part) is relevant.

1 year ago 1 0 1 0

This reminds me of all the issues with heatmaps and probes. The model really has no incentive to rely on its cot unless it is explicitly asked to do so via fine-tuning or some kind of penalty.

1 year ago 0 0 1 0

You always ask the right questions :) I don't think chain-of-thought, of current models, (except the reasoning ones) gives reliable insight about models. The issue is that cot is an output (and input) of the model, and you can change it in all sort of ways without affecting the model's output.

1 year ago 1 0 1 0

It is too early to tell :) I like the papers on your list but I think only a few of them were instant ‘classics’.

Having said that, I like: large concept models paper from meta.

1 year ago 5 0 1 0

Is the final output actually “causally” dependent on the long COT generated? How key are these traces to the search/planning clearly happening here? Some many questions but so little answers.

1 year ago 1 0 0 0
Advertisement

Great to see clarification comments. o3 is impressive nonetheless.

Played around with o1 and the ‘thinking’ Gemini model. The cot output (for Gemini) can confusing and convoluted, but it got 3/5 problems right. Stopped on the remaining 2.

These models are an impressive interpretability test bed.

1 year ago 1 0 1 0

New paper. We show that the representations of LLMs, up to 3B params(!), can be engineered to encode biophysical factors that are meaningful to experts.

We don't have to hope Adam magically finds models that learn useful features; we can optimize for models that encode for interpretable features!

1 year ago 8 1 0 0

Pinging into the void.

1 year ago 4 0 1 0