We're excited to announce the call for papers for #ICML 2026:
icml.cc/Conferences/...
See you in Seoul next summer!
Posts by Weijie Su
Great minds think alike!
Alan Turing cracked Enigma in WWII; Brad Efron asked how many words Shakespeare knew. They used the same method.
We use this method for LLM evaluation—to evaluate certain unseen capabilities of LLMs:
arxiv.org/abs/2506.02058
Another new paper that is follow-up:
arxiv.org/abs/2505.20627
It studies an alternative to RLHF: Nash learning from human feedback.
A (not so) new paper on #LLM alignment from a social choice theory viewpoint:
arxiv.org/abs/2503.10990
It reveals fundamental impossibility results concerning representing (diverse) human preferences.
Our analysis shows that it is natural to use the polar decomposition from a defining viewpoint. This gives rise to nuclear norm scaling: the update will vanish as the gradient becomes small, automatically! In contrast, Muon needs to manually tune the factor for the ortho matrix to achieve this.
We posted a paper on optimization for deep learning:
arxiv.org/abs/2505.21799
Recently there's a surge of interest in *structure-aware* optimizers: Muon, Shampoo, Soap. In this paper, we propose a unifying preconditioning perspective, offer insights into these matrix-gradient methods.
I just wrote a position paper on the relation between statistics and large language models:
Do Large Language Models (Really) Need Statistical Foundations?
arxiv.org/abs/2505.19145
Any comments are welcome. Thx!
The ranking method was tested at ICML in 2023, 2024, and 2025. I hope we'll finally use it to improve ML/AI review processes soon. Here's an article about the method, from its conception to experimentation:
www.weijie-su.com/openrank/
Our paper "The ICML 2023 Ranking Experiment: Examining Author Self-Assessment in ML/AI Peer Review" will appear in JASA as a Discussion Paper:
arxiv.org/abs/2408.13430
It's a privilege to work with such a wonderful team: Buxin, Jiayao, Natalie, Yuling, Didong, Kyunghyun, Jianqing, and Aaroth.
We're hiring a postdoc focused on the statistical foundations of large language models, starting this fall. Join our team exploring the theoretical and statistical underpinnings of LLMs. If interested, check our work: weijie-su.com/llm/ and drop me an email. #AIResearch #PostdocPosition
I wrote a post on how to connect with people (i.e., make friends) at CS conferences. These events can be intimidating so here's some suggestions on how to navigate them
I'm late for #ICLR2025 #NAACL2025, but in time for #AISTATS2025 #ICML2025! 1/3
kamathematics.wordpress.com/2025/05/01/t...
The #ICML2025 @icmlconf.bsky.social deadline has just passed!
Peer review is vital to advancing AI research. We've been conducting a survey experiment at ICML since 2023. Pls take a few minutes to participate in it, sent via email with the subject "[ICML 2025] Author Survey". Thx!
A special issue on large language models (LLMs) and statistics at Stat (onlinelibrary.wiley.com/journal/2049...). We're seeking submissions examining LLMs' impact on statistical methods, practice, education, and many more @amstatnews.bsky.social
Heading to Vancouver tomorrow for #NeurIPS2024, Dec 10-14! Excited to reconnect with colleagues and enjoy Vancouver's seafood! 🦐
Add me plz. Thx!
Machine learning has led to predictive algorithms so obscure that they resist analysis. Where does the field of traditional statistics fit into all of this? Emmanuel Candès asks the question, “Can I trust this?” Tune in to this week’s episode of “The Joy of Why” listen.quantamagazine.org/jow-321-s
Knew nothing about bluesky until today. Immediately stop using X or gradually migrate to bluesky? Is there an optimal switching strategy?