Advertisement · 728 × 90

Posts by Pietro Michiardi

LinkedIn This link will take you to a page that’s not on LinkedIn

💥 Happy to share that our paper “Information Estimation with Discrete Diffusion” has been accepted to ICLR 2026! 💥

📄 Paper (OpenReview): lnkd.in/eayzTJQb

Congratulation to our Ph.D. student Alberto Foresti, that my colleague Prof. Giulio Franzese and I jointly supervise!

2 months ago 0 0 0 0
MITHIC 2026 Unveiling Complex Interactions

Looking forward to participate in this workshop on high-order interactions and multivariate information theory:
mithic2026.github.io

3 months ago 10 3 0 0
Post image

Excited to announce a new book telling the story of mathematical approaches to studying the mind, from the origins of cognitive science to modern AI! The Laws of Thought will be published in February and is available for pre-order now.

4 months ago 167 39 2 7
Preview
TENDE: Transfer Entropy Neural Diffusion Estimation Transfer entropy measures directed information flow in time series, and it has become a fundamental quantity in applications spanning neuroscience, finance, and complex systems analysis. However, exis...

Very cool, looking forward to sit down and read your work.

Shameless plug, if you’re interested in transfer entropy estimation, you may want to have a look at our work: arxiv.org/abs/2510.14096

We’ll work more on neuroscience “applications” next, this one is rather focusing on the method.

3 months ago 3 0 0 0

With some trepidation, I'm putting this out into the world:
gershmanlab.com/textbook.html
It's a textbook called Computational Foundations of Cognitive Neuroscience, which I wrote for my class.

My hope is that this will be a living document, continuously improved as I get feedback.

3 months ago 591 238 16 10
Preview
The State Of LLMs 2025: Progress, Progress, and Predictions A 2025 review of large language models, from DeepSeek R1 and RLVR to inference-time scaling, benchmarks, architectures, and predictions for 2026.

Uploaded my State of LLMs 2025 report for this year:
magazine.sebastianraschka.com/p/state-of-l...

I planned to just write a brief overview, but yeah, it was an eventful year so it was impossible to keep it below 7000 words :D.

3 months ago 89 23 4 3
Post image

Just updated the Big LLM Architecture Comparison article...
...it grew quite a bit since the initial version in July 2025, more than doubled!
magazine.sebastianraschka.com/p/the-big-ll...

4 months ago 78 13 1 0
Advertisement
Video

Our paper introducing Alternators has been accepted at TMLR. Alternators are a new class of generative models that outperform state-space models, such as Mambas, and diffusion models such as ScoreGrad. Check it out!
Paper: arxiv.org/abs/2405.11848
Code: github.com/vertaix/Alternators
#MLSky #Vertaix

10 months ago 12 3 0 1

📢A unique fully funded three-years PhD position is open at POLITO and EURECOM.
Title: Enhancing Educational Storytelling with Human-Centered AI in the LLM Era
Application deadline: April 28, 2025
Details: www.linkedin.com/feed/update/...

11 months ago 2 4 0 0
History of Diffusion - Yang Song
History of Diffusion - Yang Song YouTube video by Bain Capital Ventures

Amazing interview with Yang Song, one of the key researchers we have to thank for diffusion models.

The most important lesson: be fearless! The community's view on score matching was quite pessimistic at the time, he went against the grain and made it work at scale!

www.youtube.com/watch?v=ud6z...

1 year ago 25 4 0 0
Preview
The State of LLM Reasoning Models Part 1: Inference-Time Compute Scaling Methods

I just shared a new article, "The State of Reasoning Models", where I am exploring 12 new research articles on improving the reasoning capabilities of LLMs (all published after the release of DeepSeek R1): magazine.sebastianraschka.com/p/state-of-l...

Happy reading!

1 year ago 61 14 1 1
Preview
Information Theoretic Text-to-Image Alignment Diffusion models for Text-to-Image (T2I) conditional generation have recently achieved tremendous success. Yet, aligning these models with user's intentions still involves a laborious trial-and-error ...

Happy to share that our paper, "Information Theoretic Text-to-Image Alignment," has been accepted at ICLR 2025! 🚀

Kudos to the team: Chao Wang, the PhD student leading this work, Alessandro Finamore, Max Gallo, and Giulio Franzese.

arxiv.org/abs/2405.20759

1 year ago 3 1 0 0
Preview
SDE Matching: Scalable and Simulation-Free Training of Latent Stochastic Differential Equations The Latent Stochastic Differential Equation (SDE) is a powerful tool for time series and sequence modeling. However, training Latent SDEs typically relies on adjoint sensitivity methods, which depend ...

Really excited about this! We note a connection between diffusion/flow models and neural/latent SDEs. We show how to use this for simulation-free learning of fully flexible SDEs. We refer to this as SDE Matching and show speed improvements of several orders of magnitude.

arxiv.org/abs/2502.02472

1 year ago 50 10 0 0
Video

How do tokens evolve as they are processed by a deep Transformer?

With José A. Carrillo, @gabrielpeyre.bsky.social and @pierreablin.bsky.social, we tackle this in our new preprint: A Unified Perspective on the Dynamics of Deep Transformers arxiv.org/abs/2501.18322

ML and PDE lovers, check it out!

1 year ago 96 16 2 0

We’re organizing once again the Gznerative Modeling Summer school this spring on the Côte d’Azur, merged for the occasion with SFdS’s Statlearn!
Applications are open!

1 year ago 10 3 0 1
Generative Modeling Summer School (GeMSS)

Excited about

🎓 Generative Modeling Summer School / Statlearn
🗓️ Mar 31 - Apr 4, 2025, France

ℹ️ gemss.ai
🎟️ Deadline: Jan 27, 2025

🧑‍🏫 Yingzhen Li, @stephanmandt.bsky.social, Aude Sportisse, Anna Korba, @glouppe.bsky.social, @jesfrellsen.bsky.social, @pamattei.bsky.social, @jmtomczak.bsky.social, TBA

1 year ago 25 11 0 1
BreimanLectureNeurIPS2024_Doucet.pdf

The slides of my NeurIPS lecture "From Diffusion Models to Schrödinger Bridges - Generative Modeling meets Optimal Transport" can be found here
drive.google.com/file/d/1eLa3...

1 year ago 326 68 9 5
Advertisement
Socrates teaching its method

Socrates teaching its method


Meet EULER (EURECOM Language Model for Education and Research) and our new #Socratic LLM for AI-Driven Education, eurecom-blog.medium.com/eurecoms-soc...
Model available @huggingface.bsky.social and @ollama.bsky.social
#LLM #RAG #EduTech

1 year ago 5 3 0 0

Great job Moto! Congrats!

1 year ago 1 0 1 0
Preview
GitHub - kuleshov-group/awesome-discrete-diffusion-models: A curated list for awesome discrete diffusion models resources. A curated list for awesome discrete diffusion models resources. - kuleshov-group/awesome-discrete-diffusion-models

Discrete diffusion has become a very hot topic again this year. Dozens of interesting ICLR submissions and some exciting attempts at scaling. Here's a bibliography on the topic from the Kuleshov group (my open office neighbors).

github.com/kuleshov-gro...

1 year ago 76 10 1 0

1/ I work in #NeuroAI, a growing field of research, which many people have only the haziest conception of...

As way of introduction to this research approach, I'll provide here a very short thread outlining the definition of the field I gave recently at our BRAIN NeuroAI workshop at the NIH.

🧠📈

1 year ago 167 48 8 12
Post image

🧠 Come work with us! In addition to our postdoc position on modelling, our lab at @institutpasteur.bsky.social we also seek a postdoc to help us create the most complete open atlas of brain folding development! This is part of our ERC Synergy project Unfold (unfold-lab.github.io). Join us! (1/10)

1 year ago 60 42 2 3
Post image

We are organising the First International Conference on Probabilistic Numerics (ProbNum 2025) at EURECOM in southern France in Sep 2025. Topics: AI, ML, Stat, Sim, and Numerics. Reposts very much appreciated!

probnum25.github.io

1 year ago 46 24 3 6