Come and work with us!
We have a PostDoc position at the intersection of ML and Biogeoscience within the TERRA excellence cluster @terra-cluster.org, w/ Senckenberg.
Be part of a great ML and Geo community and use ML to investigate fire and its impact on global vegetation๐ฅ ๐ฑ๐ณ
www.mackelab.org/jobs/
Posts by Guy Moss
This is a great opportunity to work at the intersection of ML and Biogeoscience! Based within the outstanding research community of Tรผbingen.
Reach out if you are interested!
@mackelab.bsky.social is at @cosynemeeting.bsky.social #cosyne2026 in Lisbon with two posters presented by PhD students from the lab.
Thread below on the projects ๐
On my way from Munich to Grenoble ๐ to co-lead a 3-day SBI tutorial + hackathon together with @danielged.bsky.social, organised by Pedro Rodriguez and @ugrenoblealpes.bsky.social.
Excited to meet researchers from across France, many bringing their own simulators ๐
Scientific poster with dark background and two black holes illustrated in the center. The paper visualizes gravitational waves, and explains how parameter estimation is performed with DINGO. The standard DINGO model and the DINGO-T1 architectures are illustrated and results are shown. For example, it is possible to reanalyze the same event with different detector configurations with DINGO-T1, illustrated bz a corner plot.
1/ ๐ New paper alert! We introduce Dingo-T1, a flexible transformer-based deep learning model for gravitational-wave (GW) data analysis. It adapts to different detector & frequency settings, improving inference efficiency and flexibility
๐ #AI #MachineLearning #Physics #Astronomy #AcademicSky
Finally got the job adโlooking for 2 PhD students to start spring next year:
www.gao-unit.com/join-us/
If comp neuro, ML, and AI4Neuro is your thing, or you just nerd out over brain recordings, apply!
I'm at neurips. DM me here / on the conference app or email if you want to meet ๐๏ธ๐ฎ
Iโm at NeurIPS in San Diego this week to present cool work on foundation models for SBI!
Most importantly, Iโll be around to meet people and discuss science. ๐จโ๐ฌ
Our group is at NeurIPS and EurIPS this year with four papers and one workshop poster. If you are either curious about SBI with autoML, with foundation models, or on function spaces or about differentiable simulators with Jaxley, have a look below ๐ 1/11
Iโm very grateful to my colleagues @leahsmuhle.bsky.social , @coschroeder.bsky.social, Reinhard Drews and @jakhmack.bsky.social for making this happen! Come find me at #Eurips2025 or reach out to learn more!
(9/9)
On a real geoscientific problem of inferring the surface accumulation and basal melting rates of Antarctic ice shelves, FNOPE achieves equivalent performance to previous approaches using 2 orders of magnitude fewer simulations!
(8/9)
FNOPE can be extended to estimate incredibly large parameter domains, involving >16k parameters on the Darcy flow problem.
(7/9)
Our flexible approach is amortized and can estimate posterior distributions given any discretization of the parameter and/or observation domains without any additional training.
(6/9)
FNOPE can estimate posterior distributions over 1000-dimensional parameter spaces using as few as 100 simulations on benchmark tasks.
(5/9)
This key insight leads to natural methodological extensions, such as data augmentation via masking the training data, using non-uniform Fast Fourier Transforms to work on non-uniform discretizations, and simultaneous estimation of additional vector-valued parameters.
(4/9)
Our method uses Fourier Neural Operators (FNOs) for Posterior Estimation (FNOPE). By training flow matching models with an FNO backbone, we can take into account the inductive biases of continuous parameters, forming a natural way to represent distributions over smooth functions!
(3/9)
In SBI we train generative models for posterior estimation using model simulations. However, when the parameters of interest are function-valued, we end up with very high-dimensional parameter spaces, requiring huge numbers of training simulations.
(2/9)
Iโm super excited to present our new work in #Eurips2025 and #Neurips2025! We developed FNOPE: a new simulation-based inference (SBI) method which excels at inferring function-valued parameters!
Paper: openreview.net/forum?id=yB5...
Code: github.com/mackelab/fnope
(1/9)
MackeLab has grown! ๐ Warm welcome to 5(!) brilliant and fun new PhD students / research scientists who joined our lab in the past year โ we canโt wait to do great science and already have good times together! ๐ค๐ง Meet them in the thread ๐ 1/7
Simulation-based inference (SBI) has transformed parameter inference across a wide range of domains. To help practitioners get started and make the most of these methods, we joined forces with researchers from many institutions and wrote a practical guide to SBI.
๐ Paper: arxiv.org/abs/2508.12939
๐ sbi participated in GSoC 2025 through @numfocus.bsky.social and it was a great success: our two students contributed major new features and substantial internal improvements: ๐งต ๐
Congrats to Dr Michael Deistler @deismic.bsky.social, who defended his PhD!
Michael worked on "Machine Learning for Inference in Biophysical Neuroscience Simulations", focusing on simulation-based inference and differentiable simulation.
We wish him all the best for the next chapter! ๐๐
The Macke lab is well-represented at the @bernsteinneuro.bsky.social conference in Frankfurt this year! We have lots of exciting new work to present with 7 posters (details๐) 1/9
I've been waiting some years to make this joke and now itโs real:
I conned somebody into giving me a faculty job!
Iโm starting as a W1 Tenure-Track Professor at Goethe University Frankfurt in a week (lol), in the Faculty of CS and Math
and I'm recruiting PhD students ๐ค
From hackathon to release: sbi v0.25 is here! ๐
What happens when dozens of SBI researchers and practitioners collaborate for a week? New inference methods, new documentation, lots of new embedding networks, a bridge to pyro and a bridge between flow matching and score-based methods ๐คฏ
1/7 ๐งต
Looky Looky! ๐๐ฅณ๐
arxiv.org/abs/2508.12939
Super fun project, I โค๏ธed coauthoring w/ @sbi-devs.bsky.social.
Great lead by @deismic.bsky.social & @janboelts.bsky.social. Contribs by many talented people @jakhmack.bsky.social. ๐ to #BenjaminKurtMiller for the kickstart! @helmholtzai.bsky.social
New preprint: SBI with foundation models!
Tired of training or tuning your inference network, or waiting for your simulations to finish? Our method NPE-PF can help: It provides training-free simulation-based inference, achieving competitive performance with orders of magnitude fewer simulations! โก๏ธ
I have been genuinely amazed how well tabpfn works as a density estimator, and how helpful this is for SBI ... Great work by @vetterj.bsky.social, Manuel and @danielged.bsky.social!!
My first paper on simulation-based inference (SBI) as part of @mackelab.bsky.social!
Exciting work on adapting state-of-the-art foundation models for posterior estimation. Almost plug-and-play, and surprisingly effective.
Paper/code in thread below ๐งต
New paper in Geophysical Research Letters led by Vjeran Viลกnjeviฤ mapping out ice shelf areas which are maintained by local precipitation only doi.org/10.1029/2024...
Have I been to Antarctica? No. But my colleagues have, and we can learn a lot from the data they collected! Really happy to share that our work is now published!