Advertisement ยท 728 ร— 90

Posts by Marcel Hussing

Post image Post image

Greetings from Rio, killing some time before #ICLR2026 @iclr-conf.bsky.social

1 day ago 2 0 0 0

Getting ready to fly to Rio. ๐Ÿ˜Ž๐Ÿ‡ง๐Ÿ‡ท

3 days ago 0 0 0 0

Why is the default option in the rebuttal acknowledgements at ICML to accept the paper? I'm very confused on how to use these buttons. Can someone explain them to me?

2 weeks ago 2 0 0 0

That is even if the baseline reported 80% on a similar task

3 weeks ago 3 0 0 0
Post image

Neural networks are highly non-convex, so approximate error minimizers need not look anything like each other in parameter space. But we show that nevertheless (for many model sizes) approximate error minimizers must closely agree in function/prediction space despite this!

1 month ago 9 7 1 0

It's this time of the year again: your baselines cannot be PPO and SAC.

1 month ago 2 1 1 1
Preview
How AI is changing the nature of mathematical research What machine learning theorists learned using AI agents to generate proofs โ€” and what comes next.

Michael @mkearnsphilly.bsky.social ) and I wrote a blog post about our experiences using AI for research, and our thoughts on what these developments will mean for research, publication, and education: www.amazon.science/blog/how-ai-...

1 month ago 30 13 1 3

I'm so glad that so many research problems are finally being treated as first class citizens rather than afterthoughts. ๐Ÿค”

1 month ago 1 0 0 0
Advertisement

Too many papers sound like this

Hierarchical Context-Aware Diffusion-Transformer Meta-World-Model Reinforcement Learning with Causally Disentangled Preference-Aligned Self-Supervised Compositional Multi-Scale Latent Skill Priors for Long-Horizon Generalist Decision Making

1 month ago 11 0 0 0

One reason I work on replicable and consistent RL is because it is has always been at the top of the list of criteria for reliability.

1 month ago 1 0 0 0
Post image

Excuse me? Surely telling me that didn't require much thinking.

1 month ago 1 0 2 0
Post image

I have seen multiple times now that a reviewer said sth like: the proofs are simple -> reject the paper. That is completely counter-productive. A theorem needs to generate new insights. If we learn something new from something simple that should be preferred. Don't believe me? Ask someone famous:

1 month ago 0 0 0 0

Why are you doing this to me

1 month ago 1 0 0 0

I think it's relatively simple. One side already has the job they want and the other side needs citations to get that job. And everyone tells me advertising work is how you get citations. I have been tempted to go back because on bsky, interactions have become fewer and fewer. Not going to though...

2 months ago 6 0 0 0

Yet somehow every now and then a paper becomes very popular even though its findings are similar to those of many others. This paper gets cited while the others don't. Was it just luck?

2 months ago 1 0 0 0

While I agree with the sentiment let me play devil's advocate. You only get invited to give talks if your work is already well known. Conferences have become too large to even find relevant people. And social media posts are only marginally relevant if you don't already have a large following.

2 months ago 2 0 2 0
Advertisement

I understand that that is okay but at some point it honestly becomes disheartening if you constantly have to reach out to people. There are others who don't seem to have to; what are they doing differently?

2 months ago 0 0 1 0

Iโ€™m trying to understand whether this is mostly about keyword mismatch, venue visibility, social media, etc.
For example, when I search terms like โ€œhigh update ratio RLโ€ on Scholar, our papers show up near the top.

scholar.google.com/scholar?hl=e...

Where are things going wrong?

2 months ago 1 0 1 0

Iโ€™ve been thinking about a practical question and would love some opinions:

How do your papers actually get discovered/cited?

I was searching for recent work on high update ratio RL and found several very closely related papers tackling the same failure modes we study. None cited our earlier work.

2 months ago 9 2 5 0

๐Ÿš€ Excited to share REPPO, a new on-policy RL agent!

TL;DR: Replace PPO with REPPO for fewer hyperparameter headaches and more robust training.

REPPO, led by @cvoelcker.bsky.social, will be presented at ICLR 2026. How does it work? ๐Ÿงต๐Ÿ‘‡

2 months ago 25 10 1 0
Post image Post image Post image

Scaling Laws in Particle Physics Data! This is a result I've been itching to share and it's finally out. One of the big open questions is how much better AI-based methods at particle colliders can still become. 1/4

2 months ago 16 7 3 0

Not a particle physics person but extremely curious, can you elaborate what we might hope to learn from these models in the future? What physics might we discover using them?

2 months ago 2 0 1 0

"Scientific reviewers should have experience publishing scientific work in related areas" is really not that hot of a take.

2 months ago 1 0 0 1

Clicking like on any relevant ICLR paper. Encourage people to post their work here more!

2 months ago 4 0 0 0

How do I see this?

2 months ago 2 0 1 0
Advertisement

The other paper accepted to @iclr-conf.bsky.social 2026 ๐Ÿ‡ง๐Ÿ‡ท. Our work on replicable RL sheds some light on how to consistently make decisions in RL.

@ericeaton.bsky.social @mkearnsphilly.bsky.social @aaroth.bsky.social @sikatasengupta.bsky.social @optimistsinc.bsky.social

2 months ago 13 5 0 0

Two papers accepted to @iclr-conf.bsky.social 2026! One of the is REPPO, see below! I think it deserves a lot more recognition. Let's chat about it in Rio! ๐Ÿ‡ง๐Ÿ‡ท

2 months ago 1 0 0 0

Quite disheartening that there isn't a single workshop at ICLR to present my RL work but there several topics that are listed 5 or 6 times just named differently.

2 months ago 2 0 0 0

That's correct, we did make it bold

2 months ago 1 0 0 0

Our number went down by 0.01 but it's very expensive to run so we can't have error bars. Our algorithm is so much better than the rest, new SOTA!

2 months ago 4 0 1 0