Advertisement ยท 728 ร— 90

Posts by Hadi Vafaii

Preview
Metabolic cost of information processing in Poisson variational autoencoders Computation in biological systems is fundamentally energy-constrained, yet standard theories of computation treat energy as freely available. Here, we argue that variational free energy minimization u...

Summary: we need an energy-aware theory of computation, and rate-distortion theory with Poisson latents is a good start.

But this is only the beginning, so please reach out if this sparked a thought/idea!

Here's the preprint again:
๐Ÿ“œ arXiv: arxiv.org/abs/2602.13421

[11/11] ๐Ÿงต
๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

1 month ago 2 1 1 0
Post image

The RDT framing: we can take this theory and apply it anything that can be expressed as:

๐Ÿ”นoptimization of a loss function
๐Ÿ”นapproximating conditional distributions with some variational q

=> an energy-aware approach to modeling perception, cognition, and action?

[10/11] ๐Ÿงต
๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

1 month ago 4 0 1 0
Post image

We can also motivate these results within Rate-Distortion Theory (RDT).

In RDT, you optimize a generic loss function subject to coding rate constraints.

Our contribution is that if you use Poisson latents, then:

๐Ÿ’ก coding rate budget ~ energy budget

[9/11] ๐Ÿงต

1 month ago 1 0 1 0
Post image Post image

We tested this empirically in a beta-VAE setting, and found that changing the reconstruction/KL trade-off results in a systematic different in the sparsity and metabolic cost of inference.

We saw this only in Poisson but not in Gaussian VAEs (~100% of Fristonian active inference models).

[8/11] ๐Ÿงต

1 month ago 1 0 1 0
Post image Post image

Conclusions so far:

1โƒฃ Poisson KL --> firing rate (metabolic cost)
2โƒฃ KL in general --> information rate (coding cost)

1โƒฃ + 2โƒฃ => Poisson KL couples an abstract information theoretic concept (coding rate) to a concrete biophysical quantity (firing rate)

[7/11] ๐Ÿงต

1 month ago 1 0 1 0
Post image

Now let's try to interpret this result.

First, we need to let go of a narrow way of thinking about KL divergence.

In ML, most people think KL = regularization. But it's much deeper than that:

KL = 'unique' measure of info gain (Hobson 1969: link.springer.com/article/10.1...)

[6/11] ๐Ÿงต

1 month ago 1 0 1 0
Post image Post image

What causes sparsity of representations in Poisson VAEs? The answer lies in the Poisson KL term (coding rate) in the free energy objective.

Poisson KL becomes proportional to the prior rate variable, encouraging lower neural firing rates (= lower energy consumption).

[5/11] ๐Ÿงต

1 month ago 1 0 1 0
Post image

In both ML and theoretical neuroscience, 99% of the time people default to Gaussian latents.

Previously, we showed that swapping Gaussian with Poisson results in a brain-like (spiking) generative model that reproduces sparse coding as a special case.

openreview.net/forum?id=ekt...

[4/11] ๐Ÿงต

1 month ago 1 0 1 0
Advertisement
Post image

Free energy minimization is interesting, because it unifies various theories in neuroscience with machine learning models ๐Ÿ‘‡

("Control/RL as inference" is also free energy minimization)

Different distributional and optimization choices yield different architectures.

[3/11] ๐Ÿงต

1 month ago 1 0 1 0
Post image

๐Ÿ“œ arXiv: arxiv.org/abs/2602.13421

Our starting point is the Variational Free Energy equation, which is just negative ELBO (F = -ELBO)

[2/11] ๐Ÿงต

1 month ago 2 0 1 0
Post image

The "decoupling of information and energy" is a major point of divergence between biological and artificial computers.

Energy consumption is the biggest bottleneck in scaling AI.

To address this, we need an "energy-aware theory of computation." This work is an attempt to that end.

[1/11] ๐Ÿงต

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

1 month ago 23 5 1 0
Post image

"Center for Advanced Hindsight" lmao south park predicted this in back in 2010

2 months ago 1 0 0 0

Enough #Bayesplaining. It's time to start "deriving."

We're unifying physics and neuroscience from first principles (yes, really).

Come argue with me tomorrow @ 11 AM.

@neuripsconf.bsky.social
๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

4 months ago 11 1 0 0
Post image

Here's the poster info again:

๐Ÿ“… Wed, Dec 3, 11 AM โ€” 2 PM
๐Ÿ“ Exhibit Hall C,D,E #500
๐Ÿ”— neurips.cc/virtual/2025...

See you in sunny San Diego! ๐ŸŒž

๐Ÿงต[4/4]

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

4 months ago 3 0 0 0
Post image

โœ… What's new in the paper:

We substantially improved the framing and clarity of our contributions (genuinely grateful to our amazing reviewers and their useful feedback ๐Ÿ™Œ).

+ Several new experiments (highlighted below), including scaling iP-VAE to complex color image datasets like CelebA.

๐Ÿงต[3/4]

4 months ago 4 0 1 0
Advertisement
Preview
GitHub - hadivafaii/IterativeVAE: The Official PyTorch Implementation of "Brain-like Variational Inference" (NeurIPS 2025 Paper) The Official PyTorch Implementation of "Brain-like Variational Inference" (NeurIPS 2025 Paper) - hadivafaii/IterativeVAE

Final versions of paper & code:

๐Ÿ“œ Paper: openreview.net/forum?id=573...
๐Ÿ’ป Code: github.com/hadivafaii/I...

โœ… What's new in the code:

We added a stand-alone Colab notebook implementation โ€” a great starting point if you want to understand the iP-VAE model and train your own.

๐Ÿงต[2/4]

4 months ago 3 1 1 0

I will be at @neuripsconf.bsky.social Dec 2-8 to present our "Brain-like Variational Inference" paper ๐Ÿ‘‡

Let's connect if you enjoy first-principles thinking and brain-inspired AI ๐Ÿง  ๐Ÿค–

Poster info:
๐Ÿ“… Wed, Dec 3, 11 AM โ€” 2 PM
๐Ÿ“Exhibit Hall C,D,E #500
๐Ÿ”— neurips.cc/virtual/2025...

๐Ÿงต[1/4]

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

4 months ago 19 4 1 1
Preview
Natural Gradient Works Efficiently in Learning Abstract. When a parameter space has a certain underlying structure, the ordinary gradient of a function does not represent its steepest direction, but the natural gradient does. Information geometry ...

This one is an all time favorite of mine:

direct.mit.edu/neco/article...

(tho maybe not that relevant to the "rescue" thing)

4 months ago 3 1 1 0

Thanks @hadivafaii.bsky.social for the invitation! x.com/hadivafaii/s...

4 months ago 2 1 0 0
The neuron as a direct data-driven controller (DD-DC) by Moore et al., 2024 We explored DD-DC, a provocative normative theory proposing that neurons are not just information processors, but active feedback controllers that steer their environmentโ€”including other neuronsโ€”towar...

๐Ÿ”— Meeting summary: sensorimotorai.github.io/2025/11/20/d...
๐Ÿ“œ Read the paper: www.pnas.org/doi/10.1073/...
๐Ÿ’ฌ Join the conversation on Slack: join.slack.com/t/sensorimot...

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 1 0 0 0

๐Ÿ’ก The killer insight: biological "quirks" are actually optimal control solutions:

๐Ÿ”ท STDP = solving the LQR cost (learning causal impact from feedback)
๐Ÿ”ท Noise = stabilizing "persistence of excitation"
๐Ÿ”ท ReLU = optimal policy to cross unstable fixed points

Conclusion:
โœ… Agency is cellular.

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 4 0 1 0
The neuron as a direct data-driven controller (Moore et al., 2024) โ€” Sensorimotor AI Journal Club
The neuron as a direct data-driven controller (Moore et al., 2024) โ€” Sensorimotor AI Journal Club YouTube video by Sensorimotor AI

Stop treating neurons as passive processors... They are active controllers! ๐Ÿง  ๐ŸŽฎ

In our latest Journal Club session, Thelonious Cooper presented "The neuron as a direct data-driven controller" by Moore et al., with special commentary from Mitya Chklovskii.

๐ŸŽฅ www.youtube.com/watch?v=0P7k...

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 10 1 1 0

NeurIPS 2025:

bsky.app/profile/hadi...

5 months ago 5 0 0 0
RL Debates 5: Anne โ€œnot everything is RLโ€ Collins In our 5th RL Debates presentation, Anne argued that reward-based learning is not always driven by RL computations. Sometimes itโ€™s working memory combined with outcome-insensitive habit formation, mim...

Additional links:

๐Ÿ“ Meeting summary: sensorimotorai.github.io/2025/11/13/a...
๐Ÿ“… Future meetings: sensorimotorai.github.io/schedule/
๐Ÿ’ฌ Join the conversation on Slack: join.slack.com/t/sensorimot...

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 0 0 0 0

RL Debates 5: Anne "not everything is RL" Collins

Anne delivered an amazing synthesis of her extensive work on how working memory shapes reward-based learning in humans.

๐Ÿ“œ Read the paper: nature.com/articles/s41...
๐Ÿ“ฝ๏ธ Watch the full presentation: www.youtube.com/watch?v=eEqZ...

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 7 4 1 0
Advertisement
RL Debates 4: Adam โ€œI literally measured value in the brainโ€ Lowet In our 4th RL Debates presentation, Adam presented broad topics on RL and the brain, including his distributional RL paper.

Additional links:

๐Ÿ“ Meeting summary: sensorimotorai.github.io/2025/novembe...
๐Ÿ“… Future meetings: sensorimotorai.github.io/schedule/
๐Ÿ’ฌ Join the conversation on Slack: join.slack.com/t/sensorimot...

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 1 0 0 0
RL Debates 4: Adam "I literally measured value in the brain" Lowet
RL Debates 4: Adam "I literally measured value in the brain" Lowet YouTube video by Sensorimotor AI

RL Debates 4: Adam "I literally measured value in the brain" Lowet

Adam's talk covered a lot of ground โ€” from his recent work on distributional RL (nature.com/articles/s41...) to a broader discussion of RL & the brain.

๐Ÿ“ฝ๏ธ Watch the full meeting here: www.youtube.com/watch?v=Xe7B...

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 14 3 1 0
Cover of the book โ€œthe brain abstractedโ€ by M Chirimuuta

Cover of the book โ€œthe brain abstractedโ€ by M Chirimuuta

A passage from chapter 1 including the quote

A passage from chapter 1 including the quote

Picked this one up for reading over the holidays.

โ€œThe thesis of this book is that the dominant ideas that have shaped #neuroscience are best understood as attempts to simplify the brain.โ€ ๐Ÿง 

โ€ฆwhich is itself a simplification ๐Ÿ˜‚

1 year ago 87 7 4 3
Preview
An opponent striatal circuit for distributional reinforcement learning - Nature D1- and D2-expressing striatal neurons encode separate parts of a learned reward distribution, paralleling modern approaches in machine learning.

๐Ÿ“… Next meeting: November 6, 9:00 AM PT

Where Adam Lowet will present his paper on distributional RL in brain: nature.com/articles/s41...

- See the full schedule here: sensorimotorai.github.io/schedule/
- Join our Slack: join.slack.com/t/sensorimot...

๐Ÿง ๐Ÿค–๐Ÿง ๐Ÿ“ˆ

5 months ago 1 1 0 0
Schedule Monthly journal club and RL Debate Series

๐Ÿ“œ Read the summary post: sensorimotorai.github.io/schedule/

@cortical-canonical.bsky.social @thousandbrains.org

5 months ago 1 0 1 0