Advertisement · 728 × 90

Posts by Antoine Moulin

eg, I’m always confused by the regret in advers. settings with bandit feedback. the goal is to teach a blind player to shoot a moving target, and the only reason we can say anything is bc we compare ourselves to a omniscient but paralyzed player who can only shoot in one loc. why does it make sense?

2 weeks ago 3 0 0 0

given that most regret guarantees are worst-case rather than instance-dependent I’m not sure it’s a good motivation? not that I’m aware of a good one though…

2 weeks ago 1 0 1 0

Very timely and accurate post! Why do you think of formal verification as a "short-term fix" though? If anything, it should be a necessary component of a sustainable reviewing system?

1 month ago 1 0 1 0
Preview
How AI is changing the nature of mathematical research What machine learning theorists learned using AI agents to generate proofs — and what comes next.

Michael @mkearnsphilly.bsky.social ) and I wrote a blog post about our experiences using AI for research, and our thoughts on what these developments will mean for research, publication, and education: www.amazon.science/blog/how-ai-...

1 month ago 30 13 1 3

Some text in heuristics/separation of results appear twice? Good post otherwise

4 months ago 2 0 1 0

Thank you for coming and for the great discussion!!

4 months ago 1 0 1 0
Schedule | ARLET The session will cover invited talks, contributed talks and posters. The tentative schedule in Pacific Daylight Time (GMT−7) can be found below.

So excited for the RL theory meets experiment workshop tomorrow at NeurIPS: arlet-workshop.github.io/neurips2025/...
Talks look AMAZING and you can hear me be the foolish experimentalist on a panel

4 months ago 16 1 0 1

The paper looks cool! And it seems well-written :). I did not expect these exponents in Theorem 4.2 😆

5 months ago 3 0 1 0
Advertisement
Post image

Very excited to share our preprint: Self-Speculative Masked Diffusions

We speed up sampling of masked diffusion models by ~2x by using speculative sampling and a hybrid non-causal / causal transformer

arxiv.org/abs/2510.03929

w/ @vdebortoli.bsky.social, Jiaxin Shi, @arnauddoucet.bsky.social

6 months ago 13 6 0 0
Preview
Percepta | A General Catalyst Transformation Company Transforming critical institutions using applied AI. Let's harness the frontier.

We're finally out of stealth: percepta.ai
We're a research / engineering team working together in industries like health and logistics to ship ML tools that drastically improve productivity. If you're interested in ML and RL work that matters, come join us 😀

6 months ago 118 18 7 2
Preview
Unsupervised Learning for Optimal Transport plan prediction between unbalanced graphs Optimal transport between graphs, based on Gromov-Wasserstein and other extensions, is a powerful tool for comparing and aligning graph structures. However, solving the associated non-convex optimizat...

I am happy to share that our paper "Unsupervised Learning for Optimal Transport plan prediction between unbalanced graphs" was accepted at Neurips 2025 ! 🥳

Huge thanks to my co-authors @rflamary.bsky.social and Bertrand Thirion !

arxiv.org/abs/2506.12025

(1/5)

6 months ago 31 5 1 0
Post image

News 🎉 We’re thrilled to announce our final panelist: David Silver!
Don’t miss David and our amazing lineup of speakers—submit your latest RL work to our NeurIPS workshop.
📅 Extended deadline: Sept 2 (AoE)

7 months ago 3 2 0 0
Call for Papers | ARLET A simple, whitespace theme for academics. Based on [*folio](https://github.com/bogoli/-folio) design.

We've extended the deadline for our workshop's calls for papers/ideas! Submit your work by August 29 AoE. Instructions on the website: arlet-workshop.github.io/neurips2025/...

8 months ago 3 2 0 0
Post image

The OpenReview link for our calls (for papers and ideas) is available, submit here: openreview.net/group?id=Neu...

We look forward to receiving your submissions!

8 months ago 2 1 0 0

last year's edition was so much fun I'm really looking forward to this one!! join us in San Diego :))

8 months ago 4 0 0 0

Was it recorded? 🤔

10 months ago 1 0 1 0
Advertisement
Post image

Join us for Nneka's presentation tomorrow! Last talk before the summer break.

10 months ago 9 3 0 0
Post image

Join us tomorrow for Dave's talk! He will present his recent work on randomised exploration, which received an outstanding paper award at ALT 2025 earlier this year.

10 months ago 4 1 0 0
Preview
Inverse Q-Learning Done Right: Offline Imitation Learning in $Q^π$-Realizable MDPs We study the problem of offline imitation learning in Markov decision processes (MDPs), where the goal is to learn a well-performing policy given a dataset of state-action pairs generated by an expert...

link: arxiv.org/abs/2505.19946

10 months ago 3 0 0 0
Post image Post image

new preprint with the amazing @lviano.bsky.social and @neu-rips.bsky.social on offline imitation learning! learned a lot :)

when the expert is hard to represent but the environment is simple, estimating a Q-value rather than the expert directly may be beneficial. lots of open questions left though!

10 months ago 18 3 1 1
Preview
Nonlinear Meta-Learning Can Guarantee Faster Rates Many recent theoretical works on \emph{meta-learning} aim to achieve guarantees in leveraging similar representational structures from related tasks towards simplifying a target task. The main aim of ...

🚨 New paper accepted at SIMODS! 🚨
“Nonlinear Meta-learning Can Guarantee Faster Rates”

arxiv.org/abs/2307.10870

When does meta learning work? Spoiler: generalise to new tasks by overfitting on your training tasks!

Here is why:
🧵👇

10 months ago 9 7 2 1

Dhruv Rohatgi will be giving a lecture on our recent work on comp-stat tradeoffs in next-token prediction at the RL Theory virtual seminar series (rl-theory.bsky.social) tomorrow at 2pm EST! Should be a fun talk---come check it out!!

10 months ago 11 5 1 0
Video

new work on computing distances between stochastic processes ***based on sample paths only***! we can now:
- learn distances between Markov chains
- extract "encoder-decoder" pairs for representation learning
- with sample- and computational-complexity guarantees
read on for some quick details..
1/n

10 months ago 37 10 1 0
Preview
Discrete Diffusion: Continuous-Time Markov Chains A tutorial explaining some key intuitions behind continuous time Markov chains for machine learners interested in discrete diffusion models: alternative representations, connections to point processes...

A new blog post with intuitions behind continuous-time Markov chains, a building block of diffusion language models, like @inceptionlabs.bsky.social's Mercury and Gemini Diffusion. This post touches on different ways of looking at Markov chains, connections to point processes, and more.

10 months ago 21 5 1 0
Advertisement

oh the inference blog posts are back 🥰

10 months ago 0 0 1 0

Mattes Mollenhauer, Nicole M\"ucke, Dimitri Meunier, Arthur Gretton: Regularized least squares learning with heavy-tailed noise is minimax optimal https://arxiv.org/abs/2505.14214 https://arxiv.org/pdf/2505.14214 https://arxiv.org/html/2505.14214

11 months ago 6 6 1 1
Post image

Later today, Sikata and Marcel will talk about their recent work on oracle-efficient RL with ensembles. Join us!

11 months ago 6 4 0 0
Video

Excited to share what I've been up to: bringing text diffusion to Gemini!

Diffusion models are _fast_, and hold immense promise to challenge autoregressive models as the de facto standard for language modeling.

11 months ago 15 2 1 0

omg thanks

11 months ago 0 0 0 0
Community events and tutorials, list from the website

Community events and tutorials, list from the website

Workshops, list from the website

Workshops, list from the website

The tutorials, workshops, and community events for #COLT2025 have been announced!

Exciting topics, and impressive slate of speakers and events, on June 30! The workshops have calls for contributions (⏰ May 16, 19, and 25): check them out!
learningtheory.org/colt2025/ind...

11 months ago 20 7 2 0