Advertisement · 728 × 90

Posts by Herbie(Zizhan) He

Preview
Building spatial world models from sparse transitional episodic memories Many animals possess a remarkable capacity to rapidly construct flexible cognitive maps of their environments. These maps are crucial for ethologically relevant behaviors such as navigation, explorati...

[8/8] We believe ESWM points to a new generation of brain-inspired models—ones that reason over fragments, generalize across structure, and adapt efficiently to change.

👥W/ @maximemdaigle.bsky.social, @bashivan.bsky.social

Read the full paper: arxiv.org/abs/2505.13696

1 month ago 1 0 0 0
Post image

[7/8] Beyond Grid World, ESWM is scalable to the more complex MiniGrid (high-dimensional observation) and 3D indoor scenes ProcThor (realistic pixel observations).

1 month ago 1 0 1 0
Post image

[6/8] When environments change (e.g., new obstacles), ESWM adapts by updating its temporally and spatially independent memories. No retraining is needed.

1 month ago 1 0 1 0
Post image

[5/8] ESWM also supports efficient exploration by acting on uncertainty to collect experiences and navigate between states.

1 month ago 1 0 1 0
Post image Post image Post image

[4/8] In GridWorld experiments: 1) Transformer >> LSTM & Mamba. 2) ESWM generalizes to novel observations and structures. 3) Its latent space reflects the environment structure. 4) It predicts by integrating independent transitions.

1 month ago 1 1 1 0
Post image Post image Post image

[3/8] ESWM is designed to operate on sets of temporally independent transitions. Given such a set, it infers unseen transitions. The model is meta-trained across environments to support generalization. We show three settings in which we validate ESWM.

1 month ago 1 0 1 0
Post image

[2/8] In contrast, neuroscience evidence suggests that animals can build spatial representation across independent experiences (i.e day1: A->B, day2: B->C, day3: infers A->C). Motivated by these observations, we introduce ESWM:

1 month ago 2 0 1 0

[1/8] Existing world models rely on a sequence of observations to predict future states. This leads to: 1) redundancy due to temporal overlap (contexts grow for large envs), 2) limited adaptability when environments change due to temporal dependency.

1 month ago 2 0 1 0
Advertisement
Video

Video abstract [2/2]

1 month ago 1 0 1 0
Video

New paper 🚨 #ICLR26

Most world models predict the future from a past trajectory. But neuroscience suggests that such inference can instead be made from temporally independent experiences.

We built the Episodic Spatial World Model (ESWM), a model that does exactly this:

Video abstract [1/2]

1 month ago 7 3 1 1
Post image Post image

New paper 🚨
"Stable Deep Reinforcement Learning via Isotropic Gaussian Representations"

Deep RL suffers from unstable training, representation collapse, and neuron dormancy. We show that a simple geometric insight, isotropic Gaussian representations, can fix this. Here's how 👇

1 month ago 24 4 2 2
Preview
Building spatial world models from sparse transitional episodic memories Many animals possess a remarkable capacity to rapidly construct flexible mental models of their environments. These world models are crucial for ethologically relevant behaviors such as navigation, ex...

[12/12]
We believe ESWM points to a new generation of brain-inspired models—ones that reason over fragments, generalize across structure, and adapt on the fly.

📄 arxiv.org/abs/2505.13696
👥 @maximemdaigle.bsky.social , @bashivan.bsky.social

9 months ago 2 0 0 0
Post image

[11/12]
🔧 When environments change—say a new wall appears—ESWM adapts instantly. No retraining is needed. Just update the memory bank and the model replans.

This separation of memory and reasoning makes ESWM highly flexible.

9 months ago 2 0 1 0
Post image

[10/12]
🧭 It gets even better!

ESWM can navigate between arbitrary points using only its memory bank—planning efficiently in latent space with near-optimal paths.

No access to global maps or coordinates required.

9 months ago 2 0 1 0
Post image

[9/12]
🚶 With no additional training, ESWM can explore novel environments efficiently by acting on uncertainty.

9 months ago 2 0 1 0
Advertisement
Post image

[8/12]
⚙️ How are these maps built?

We find that ESWM stitches together memories via overlapping states—merging local transitions into global structure.

Obstacles and boundaries serve as spatial anchors, guiding how memories are organized in latent space.

9 months ago 3 1 2 0
Post image

[7/12]
🏞️ How does ESWM solve the task?

Using ISOMAP, we visualize its latent representations—beautifully organized spatial layouts emerge from its internal states, even when the model sees only a small part or out-of-distribution environments.

9 months ago 2 0 1 0
Post image

[6/12]
⚡️ Transformer-based ESWM models outperform LSTMs and Mamba, especially in settings where observations are compositional. Attention allows the model to flexibly bind relevant memories and generalize across structures.

9 months ago 2 0 1 0
Post image Post image

[5/12]
To train ESWM, we use meta-learning across diverse environments. At test time, the model gets a minimal set of disjoint episodic memories (single transitions) and must predict a missing element in a new transition—without ever seeing the full map.

9 months ago 2 0 1 0
Post image

[4/12]
🧠 Inspired by the MTL’s architecture and function, we built ESWM: a neural network that infers the structure of its environment from isolated, one-step transitions—just like the brain integrates episodes into a cognitive map.

9 months ago 2 1 1 0

[3/12]
Such shortcut-seeking behavior is supported by internal world models in the brain’s medial temporal lobe (MTL), an area involved in both episodic memory and spatial navigation.

9 months ago 2 0 1 0
Preview
Rats take correct novel routes and shortcuts in an enclosed maze - PubMed In 3 experiments, rats were allowed to travel selected routes along the internal alleys of a cross-maze that led from one distinctive end box to another. The maze and procedures used were designed to ...

[2/12]
Many animals can infer unexperienced paths by integrating disjoint memories. Mice, e.g., take shortcuts they’ve never physically traversed: pubmed.ncbi.nlm.nih.gov/17469957/

9 months ago 2 0 1 0
Post image

🧠 Can a neural network build a spatial map from scattered episodic experiences like humans do?

We introduce the Episodic Spatial World Model (ESWM)—a model that constructs flexible internal world models from sparse, disjoint memories.

🧵👇 [1/12]

9 months ago 20 2 3 2