Advertisement · 728 × 90

Posts by Eva Yi Xie

Preview
Neuroscience has a species problem If neuroscience is serious about building general principles of brain function, cross-species dialogue must become a core organizing principle.

Differences between species should be treated as informative constraints that refine theory, not as inconsistencies to be explained away, writes @suthanalab.bsky.social.

#neuroskyence

www.thetransmitter.org/animal-model...

2 months ago 22 9 0 1

🤠 We are in Upper Level Room 10!

4 months ago 1 0 0 0
Post image

2️⃣ days until #NeurIPS2025 Data on the Brain & Mind workshop! 🧠💭🤖 Join us on Dec 7 for a full-day interactive session 8am-5pm PT.

Authors, please remember to RSVP for our mentorship lunch 🥙 generously supported by @kavlifoundation.org and @simonsfoundation.org (@flatironinstitute.org)

4 months ago 0 1 0 1

It’s TODAY! 🤗
📍 Exhibit Hall C,D,E #2109
⏰ 4:30-7:30pm

4 months ago 1 0 0 0

🗓️ Our #NeurIPS2025 poster is in 2 days on Dec 4! “Slow Transition to Low-Dimensional Chaos in Heavy-Tailed Recurrent Neural Networks”
📍 Exhibit Hall C,D,E #2109
⏰ 4:30-7:30pm
We will also be at NeurReps workshop on Dec 7 🙌 Plz come say hi! Happy to chat :)

(I’ll also be at UniReps + DBM workshops)

4 months ago 1 0 0 0

Thank you 😊 That means a lot to hear!

5 months ago 0 0 0 0

#neuroscience

5 months ago 1 0 0 0

9/9 Lastly, we thank the colleagues @alleninstitute.org and @cosynemeeting.bsky.social for their insightful feedback on an early version of this work! Happy to chat: evayixie@princeton.edu; lukasz.kusmierz@alleninstitute.org.

5 months ago 1 0 1 0
Advertisement

8/ @tyrellturing.bsky.social ‘s group recently shows brain-like learning with exponentiated gradients naturally gives rise to log-normal connectivity distributions—our results offer a theoretical perspective that elucidates the dynamical consequences of these heavy-tailed structures.

5 months ago 1 0 1 0
Slow Transition to Low-Dimensional Chaos in Heavy-Tailed Recurrent... Growing evidence suggests that synaptic weights in the brain follow heavy-tailed distributions, yet most theoretical analyses of recurrent neural networks (RNNs) assume Gaussian connectivity. We...

7/ For more details, implications of our results to neuroscience 🧠 and machine learning 🤖, + exciting future directions, please check out our full paper or visit our poster at #NeurIPS2025:

🔗OpenReview: openreview.net/forum?id=J0S...
📍Code: github.com/AllenInstitu...

5 months ago 2 0 1 0

6/ Conclusion: Our results reveal a biologically aligned tradeoff between the robustness of dynamics and the richness of neural activity. Our results provide a tractable framework for understanding dynamics in realistically sized, heavy-tailed neural circuits.

5 months ago 1 0 1 0
Post image

5/ ‼️Result 3: However, this robustness of slow transition comes with a tradeoff ↔️: heavier tails reduce the Lyapunov dimension of the network attractor, indicating lower effective dimensionality.

5 months ago 2 0 1 0

4/ (Side note: The computational benefit of being near the edge of chaos is well established for both feedforward and recurrent neural networks. We validate in Appendix L this indeed translates to improved info processing in simple reservoir-computing tasks. 🤖🧠)

5 months ago 3 0 1 0
Post image

3/ 🔎Result 2: Compared to Gaussian networks, we found finite heavy-tailed RNNs exhibit a broader gain regime near the edge of chaos: a *slow* transition to chaos. 🐢

5 months ago 2 0 1 0
Post image

2/ 🔎Result 1: While mean-field theory for the infinite system predicts ubiquitous chaos, our analysis reveals *finite-size* RNNs have a sharp transition between quiescent & chaotic dynamics. 

We theoretically predict the gain of transition and validated it through simulations.

5 months ago 3 0 1 0

1/ Setup: With @mihalas.bsky.social and Lukasz Kusmierz, We study RNNs with weights drawn from biologically plausible Lévy alpha-stable distributions, generalizing the Gaussian distribution to heavy tails.

5 months ago 1 0 1 0
https://tinyurl.com/heavyrnn

https://tinyurl.com/heavyrnn

Connectome suggests brain’s synaptic weights follow heavy-tailed distributions, yet most analyses of RNNs assume Gaussian connectivity. 

🧵⬇️ Our @alleninstitute.org #NeurIPS2025 paper shows heavy-tailed weights can strongly affect dynamics, trade off robustness + attractor dimension.

5 months ago 31 7 3 2
Post image

🚨 Only 4 days left to submit to Data on the Brain and Mind! 🚨

Don’t miss your chance to contribute to our Findings or Tutorial tracks.
We’re excited to feature oral presentations in both tracks!

7 months ago 2 2 0 1
Advertisement
NeurIPS 2025 Workshop DBM Welcome to the OpenReview homepage for NeurIPS 2025 Workshop DBM

🚨 Deadline Extended 🚨
The submission deadline for the Data on the Brain & Mind Workshop (NeurIPS 2025) has been extended to Sep 8 (AoE)! 🧠✨
We invite you to submit your findings or tutorials via the OpenReview portal:
openreview.net/group?id=Neu...

7 months ago 4 2 0 0
Data on the Brain & Mind

📢 10 days left to submit to the Data on the Brain & Mind Workshop at #NeurIPS2025!

📝 Call for:
• Findings (4 or 8 pages)
• Tutorials

If you’re submitting to ICLR or NeurIPS, consider submitting here too—and highlight how to use a cog neuro dataset in our tutorial track!
🔗 data-brain-mind.github.io

7 months ago 8 5 0 0
Post image

🚨 Excited to announce our #NeurIPS2025 Workshop: Data on the Brain & Mind

📣 Call for: Findings (4- or 8-page) + Tutorials tracks

🎙️ Speakers include @dyamins.bsky.social @lauragwilliams.bsky.social @cpehlevan.bsky.social

🌐 Learn more: data-brain-mind.github.io

8 months ago 31 10 0 3