When and how can test-time thinking allow models to use information latent in their training data? What are the benefits and tradeoffs relative to other solutions like synthetic data augmentation? Pleased to share (after a long delay) an exploration of these issues: arxiv.org/abs/2604.01430 thread:
Posts by Taylor Webb
The brain can clearly represent novel combinations of features, we need to be able to explain how it does it!
potentially be very important. Many would argue that the ability to adaptively handle these types of situations is a very important part of human intelligence. You seem to be arguing that these cases are rare and therefore uninteresting or not important but that doesn’t follow.
Is there any data to support the claim that precomputed conjunctions can handle the vast majority of cases? Even accepting that novel conjunctions are rare (which seems unlikely to me) that doesn’t mean they’re unimportant to the survival of an organism. Novel but rare situations can…
To extend your lunch metaphor, sometimes you have to make your own lunch!
same basic computation operating over different timescales. Precomputed conjunctive codes represent stable statistical regularities, which can get you pretty far, but sometimes you need to represent novel conjunctions for which you don’t already have precomputed codes.
associative learning as implemented with hopfield networks, so one way to think about this is that the models are performing the same kind of associative learning that encodes natural statistics on a longer timescale, but here over a short timescale. I’m not sure why you’re opposed to this…
separate tokens with shared parameters. Without this architectural feature, ie if position and feature embeddings for separate objects were additively combined in a single vector space, the assignment (binding!) of features to objects would be lost. Self attention is isomorphic to…
the models handle this by using separate tokens to store particular feature conjunctions. When objects span multiple tokens, position ID embeddings track the assignment of tokens to objects, ie they bind those tokens together. This solution fundamentally depends on the presence of…
Yes how the system handles the ‘exceptions’ is precisely the question. Dynamic reweighting and attention alone don’t accomplish that. You need some way to encode the in-context conjunction of features, ie binding! As we show in the paper (and other recent work has shown in other domains)…
The study from Li et al and other recent studies eg arxiv.org/html/2506.15... show that vision models develop emergent index-like representations. These results illustrate how these models solve the binding problem, not that binding was unnecessary in the first place.
You claim that precomputed conjunctive codes can support 99% of visual processing. Regardless of whether this is the right number, what about the remaining 1%? People can obviously perceive synthetic scenes with arbitrary feature combinations, how?
A new Department of Cognitive Science is being created at Bocconi University in Milan, Italy.
Here is the call for a cluster hire (for around 10 faculty) in all areas of cognitive science, at both junior and senior levels:
www.unibocconi.it/en/faculty-a...
Deadline: May 4th, 2026
Excited to give a talk at #Cosyne2026 about my PhD work!
We show that RNNs trained on visual search converge on brain-like solutions, producing primate-like behavior and neural representations. Happy to chat if you're at Cosyne!
📅 March 15, 2026
📍 Lisbon, Portugal
www.biorxiv.org/content/10.1...
Can large language models *introspect*?
In a new paper, @kmahowald.bsky.social and I study the MECHANISM of introspection in big open-source models.
tldr: Models detect internal anomalies through DIRECT ACCESS, but don't know what the anomalies are.
And they love to guess “apple” 🍎
I like this general point about levels of explanation (which I think is similar to the point we make here arxiv.org/abs/2508.05776) but how does it relate to the discussion about mistakes made by LLMs? (Possibly explained by the earlier context of the clip)
But this example again appears not to involve the reasoning mode. I agree that ‘thinking’ is confusing nomenclature but it’s notable that most (not all) of the stupidest mistakes come from the feedforward / parallel processing mode.
We also have a paper on this arxiv.org/abs/2411.00238 but it doesn’t seem to be an arbitrary failure. Instead they seem to fail in precisely the ways that human vision fails under time pressure (including with counting), and increasingly the models seem to resolve this via sequential processing.
Re: seeing vs. thinking, 'thinking' is arguably a bad term for this, in the vision setting the models depend on sequential processing to individuate objects but that wouldn't commonly be referred to as 'thinking' in the colloquial sense.
Huh, seems to be the result of different prompts, although it's arguably confusing to say that no liquid can be poured into it (at all or only in the current configuration?). In general most of the comically stupid mistakes (e.g. how many b's in blueberry) seem to be from the non-thinking models.
Funny, but these demos always seem to be the free / instant model. With thinking turned on it gets this correct.
very interesting! Consistent with this, we found that induction heads seem to be completely distinct from what we called 'symbolic induction heads' i.e. function vector heads arxiv.org/abs/2502.20332
How do you knock the induction heads out of an LM while preserving its ability to think? Is it even possible?
@keremsahin22.bsky.social's work is worth reading if you haven't seen it yet.
hapax.baulab.info
What is the relationship between memorization and generalization in AI? Is there a fundamental tradeoff? In infinitefaculty.substack.com/p/memorizati... I’ve reviewed some of the evolving perspectives on memorization & generalization in machine learning, from classic perspectives through LLMs.
Unfortunately the event is in-person only.
Very excited for our second workshop on the computational ingredients of reasoning (Feb 24-27), this one focused on mechanisms of reasoning in both AI and the brain. Check out the program to see our amazing lineup of speakers, and please consider attending! ivado.ca/en/events/me...
Thrilled that my paper is out in the @nature.com. We explored how the brain builds complex tasks by compositionally combining simpler sub-task representations. The brain flexibly performs multiple tasks by dynamically reusing neural subspaces for sensory inputs and motor actions
rdcu.be/eRVUk
Excited to announce a new book telling the story of mathematical approaches to studying the mind, from the origins of cognitive science to modern AI! The Laws of Thought will be published in February and is available for pre-order now.
That is, in order to do the kinds of things that are supposed to require algebraic / rule-based operations, these models actually do something that is algebraic, which both affirms the importance of algebraic operations for human-like reasoning and also shows it doesn't need to be innate.
I generally agree, but the interesting thing is that the LLMs/VLMs sometimes do end up doing something very structured and algebraic, as we show e.g. here arxiv.org/abs/2502.20332 and here arxiv.org/abs/2506.15871 (the paper that @romanfeiman.bsky.social 's meme was commenting on).