Summary: we need an energy-aware theory of computation, and rate-distortion theory with Poisson latents is a good start.
But this is only the beginning, so please reach out if this sparked a thought/idea!
Here's the preprint again:
๐ arXiv: arxiv.org/abs/2602.13421
[11/11] ๐งต
๐ง ๐ค๐ง ๐
Posts by Hadi Vafaii
The RDT framing: we can take this theory and apply it anything that can be expressed as:
๐นoptimization of a loss function
๐นapproximating conditional distributions with some variational q
=> an energy-aware approach to modeling perception, cognition, and action?
[10/11] ๐งต
๐ง ๐ค๐ง ๐
We can also motivate these results within Rate-Distortion Theory (RDT).
In RDT, you optimize a generic loss function subject to coding rate constraints.
Our contribution is that if you use Poisson latents, then:
๐ก coding rate budget ~ energy budget
[9/11] ๐งต
We tested this empirically in a beta-VAE setting, and found that changing the reconstruction/KL trade-off results in a systematic different in the sparsity and metabolic cost of inference.
We saw this only in Poisson but not in Gaussian VAEs (~100% of Fristonian active inference models).
[8/11] ๐งต
Conclusions so far:
1โฃ Poisson KL --> firing rate (metabolic cost)
2โฃ KL in general --> information rate (coding cost)
1โฃ + 2โฃ => Poisson KL couples an abstract information theoretic concept (coding rate) to a concrete biophysical quantity (firing rate)
[7/11] ๐งต
Now let's try to interpret this result.
First, we need to let go of a narrow way of thinking about KL divergence.
In ML, most people think KL = regularization. But it's much deeper than that:
KL = 'unique' measure of info gain (Hobson 1969: link.springer.com/article/10.1...)
[6/11] ๐งต
What causes sparsity of representations in Poisson VAEs? The answer lies in the Poisson KL term (coding rate) in the free energy objective.
Poisson KL becomes proportional to the prior rate variable, encouraging lower neural firing rates (= lower energy consumption).
[5/11] ๐งต
In both ML and theoretical neuroscience, 99% of the time people default to Gaussian latents.
Previously, we showed that swapping Gaussian with Poisson results in a brain-like (spiking) generative model that reproduces sparse coding as a special case.
openreview.net/forum?id=ekt...
[4/11] ๐งต
Free energy minimization is interesting, because it unifies various theories in neuroscience with machine learning models ๐
("Control/RL as inference" is also free energy minimization)
Different distributional and optimization choices yield different architectures.
[3/11] ๐งต
๐ arXiv: arxiv.org/abs/2602.13421
Our starting point is the Variational Free Energy equation, which is just negative ELBO (F = -ELBO)
[2/11] ๐งต
The "decoupling of information and energy" is a major point of divergence between biological and artificial computers.
Energy consumption is the biggest bottleneck in scaling AI.
To address this, we need an "energy-aware theory of computation." This work is an attempt to that end.
[1/11] ๐งต
๐ง ๐ค๐ง ๐
"Center for Advanced Hindsight" lmao south park predicted this in back in 2010
Enough #Bayesplaining. It's time to start "deriving."
We're unifying physics and neuroscience from first principles (yes, really).
Come argue with me tomorrow @ 11 AM.
@neuripsconf.bsky.social
๐ง ๐ค๐ง ๐
Here's the poster info again:
๐
Wed, Dec 3, 11 AM โ 2 PM
๐ Exhibit Hall C,D,E #500
๐ neurips.cc/virtual/2025...
See you in sunny San Diego! ๐
๐งต[4/4]
๐ง ๐ค๐ง ๐
โ
What's new in the paper:
We substantially improved the framing and clarity of our contributions (genuinely grateful to our amazing reviewers and their useful feedback ๐).
+ Several new experiments (highlighted below), including scaling iP-VAE to complex color image datasets like CelebA.
๐งต[3/4]
Final versions of paper & code:
๐ Paper: openreview.net/forum?id=573...
๐ป Code: github.com/hadivafaii/I...
โ
What's new in the code:
We added a stand-alone Colab notebook implementation โ a great starting point if you want to understand the iP-VAE model and train your own.
๐งต[2/4]
I will be at @neuripsconf.bsky.social Dec 2-8 to present our "Brain-like Variational Inference" paper ๐
Let's connect if you enjoy first-principles thinking and brain-inspired AI ๐ง ๐ค
Poster info:
๐
Wed, Dec 3, 11 AM โ 2 PM
๐Exhibit Hall C,D,E #500
๐ neurips.cc/virtual/2025...
๐งต[1/4]
๐ง ๐ค๐ง ๐
This one is an all time favorite of mine:
direct.mit.edu/neco/article...
(tho maybe not that relevant to the "rescue" thing)
Thanks @hadivafaii.bsky.social for the invitation! x.com/hadivafaii/s...
๐ Meeting summary: sensorimotorai.github.io/2025/11/20/d...
๐ Read the paper: www.pnas.org/doi/10.1073/...
๐ฌ Join the conversation on Slack: join.slack.com/t/sensorimot...
๐ง ๐ค๐ง ๐
๐ก The killer insight: biological "quirks" are actually optimal control solutions:
๐ท STDP = solving the LQR cost (learning causal impact from feedback)
๐ท Noise = stabilizing "persistence of excitation"
๐ท ReLU = optimal policy to cross unstable fixed points
Conclusion:
โ
Agency is cellular.
๐ง ๐ค๐ง ๐
Stop treating neurons as passive processors... They are active controllers! ๐ง ๐ฎ
In our latest Journal Club session, Thelonious Cooper presented "The neuron as a direct data-driven controller" by Moore et al., with special commentary from Mitya Chklovskii.
๐ฅ www.youtube.com/watch?v=0P7k...
๐ง ๐ค๐ง ๐
NeurIPS 2025:
bsky.app/profile/hadi...
Additional links:
๐ Meeting summary: sensorimotorai.github.io/2025/11/13/a...
๐
Future meetings: sensorimotorai.github.io/schedule/
๐ฌ Join the conversation on Slack: join.slack.com/t/sensorimot...
๐ง ๐ค๐ง ๐
RL Debates 5: Anne "not everything is RL" Collins
Anne delivered an amazing synthesis of her extensive work on how working memory shapes reward-based learning in humans.
๐ Read the paper: nature.com/articles/s41...
๐ฝ๏ธ Watch the full presentation: www.youtube.com/watch?v=eEqZ...
๐ง ๐ค๐ง ๐
Additional links:
๐ Meeting summary: sensorimotorai.github.io/2025/novembe...
๐
Future meetings: sensorimotorai.github.io/schedule/
๐ฌ Join the conversation on Slack: join.slack.com/t/sensorimot...
๐ง ๐ค๐ง ๐
RL Debates 4: Adam "I literally measured value in the brain" Lowet
Adam's talk covered a lot of ground โ from his recent work on distributional RL (nature.com/articles/s41...) to a broader discussion of RL & the brain.
๐ฝ๏ธ Watch the full meeting here: www.youtube.com/watch?v=Xe7B...
๐ง ๐ค๐ง ๐
Cover of the book โthe brain abstractedโ by M Chirimuuta
A passage from chapter 1 including the quote
Picked this one up for reading over the holidays.
โThe thesis of this book is that the dominant ideas that have shaped #neuroscience are best understood as attempts to simplify the brain.โ ๐ง
โฆwhich is itself a simplification ๐