Thank you!
Posts by Alexander Terenin
Thanks! It may look a certain way from outside, but boldness here is a product of conviction and constraints - not one, or the other, but both. Just me doing the best I can to make things happen, however has the chance of working.
Thank you!
Big career news: I'm leaving academia - and moving to the San Francisco Bay Area to explore something new.
I've written a short blog post with a few reflections on the end of this chapter.
If you'd like to catch up, now is the time to reach out!
avt.im/blog/the-roa...
Happy to share a major milestone: after years of development, we are officially launching Version 1.0 of the GeometricKernels library!
To top it off, our accompanying paper has just been published in JMLR (MLOSS)! 🎉
github.com/geometric-ke...
I wrote a long, detailed blog post on the state of AI and machine learning.
The post's purpose is to sharpen my thinking and help ensure I work on the right things over the next few years.
You might find parts of it interesting. Comments are welcome.
avt.im/blog/where-a...
After a delayed flight and missing my first poster, I have officially arrived at NeurIPS!
I’ll present another poster tomorrow at CDE 606 from 11-2. I’ll post more on this soon.
If you’re interested in meeting up, let’s get in touch!
Today, I gave a talk at the INFORMS Job Market Showcase!
If you're interested, here are the slides - link below!
presentations.avt.im/2025-10-26-A...
I am hiring a fully-funded #PhD in #ML to work at the University of Edinburgh on 𝐠𝐞𝐨𝐦𝐞𝐭𝐫𝐢𝐜 𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 and 𝐮𝐧𝐜𝐞𝐫𝐭𝐚𝐢𝐧𝐭𝐲 𝐪𝐮𝐚𝐧𝐭𝐢𝐟𝐢𝐜𝐚𝐭𝐢𝐨𝐧.
Application deadline: 31 Dec '25. Starts May/Sep '26.
Details in the reply.
Pls RT and share with anyone interested!
Check out the work at: arxiv.org/abs/2502.14790
And, again, shoutout to amazing coauthor Jeff Negrea! Working together has been a great pleasure!
Stay tuned for follow-up: we've been working on using this viewpoint to understand other correlated perturbation-based algorithms.
So this sums up the work! If you followed along, thanks for the interest!
I think you'd agree that "Bayesian Algorithms for Adversarial Online Learning: from Finite to Infinite Action Spaces" is a much better title than before. The old one was much harder to pronounce.
The Bayesian viewpoint proves useful for developing this analysis.
It allows us to guess what a good prior will be, and suggests ways to use probability as a tool to prove the algorithm works.
We prove that the Bayesian approach works in this setting too.
To achieve this, we develop a new probabilistic analysis of correlated Gaussian follow-the-perturbed-leader algorithms, of which ours is a special case.
This has been an open challenge in the area.
The second one is where X = [0,1]^d and Y is the space of bounded Lipschitz functions.
Here, you can't use a prior with independence across actions. You need to share information between actions.
We do this by using a Gaussian process, with correlations between actions.
The first one is the classical discrete setting where standard algorithms such as exponential weights are studied.
You can use a Gaussian prior which is independent across actions.
Okay, so we now know what "Bayesian Algorithms for Adversarial Online Learning" are.
What about "from Finite to Infinite Action Spaces"?
This covers the two settings we show the aforementioned results in.
This approach appears to not make any sense: the Bayesian model is completely fake.
We're pretending to know a distribution for how the adversary will act in the future.
But, in reality, they can do anything.
And yet... we show that this works!
We show that this game secretly has a natural Bayesian strategy - one we show is strong.
What's the strategy?
It's really simple:
- Place a prior distribution of what the adversary will do in the future
- Condition on what the adversary has done
- Sample from the posterior
One observation about adversarial online learning is that it appears to have nothing to do with Bayesian learning.
There is a two-player zero-sum game, not a joint probability distribution.
So you can't just solve it by applying Bayes' Rule. Or can you?
Okay, so now we understand what "Adversarial Online Learning" is.
We propose "Bayesian Algorithms" for this.
What does that mean? Let's unpack.
Online learning is therefore a good model for learning to explore by taking random actions.
In contrast to other approaches to resolving explore-exploit tradeoffs such as upper confidence bounds which produce purely deterministic strategies.
So that's our setting. Why's it interesting?
Because many other hard decision problems can be reduced to online learning, including certain forms of reinforcement learning (via decision-estimation coefficients), equilibrium computation (via no-regret dynamics), and others.
Specifically, their goal is to minimize regret
R(p,q) = E_{x_t~p_t, y_t~q_t} \sup_{x\in X} \sum_{t=1}^T y_t(x) - \sum_{t=1}^T y_t(x_t).
Meaning, the learner compares the sum of their rewards y_t(x_t) with the sum of y_t(x) for the best possible single non-time-dependent x.
The learner's goal is to achieve the highest rewards possible. But at each time, the adversary can choose a different reward function.
So why is this game not impossible?
Because the learner only compares how well they do with the *sum* of the adversary's previous rewards.
"Adversarial Online Learning" refers to a two-player zero-sum repeated game between a learner and adversary.
At each time point:
- The learner chooses a distribution of predictions p_t over an action space X.
- The adversary chooses a reward function y_t : X -> R.
First a link: arxiv.org/abs/2502.14790
Now, let's unpack the new title!
Let's start with what we mean by "Adversarial Online Learning".
Paper update: our recent work on Thompson sampling has a shiny new - and I hope much better - name!
This new name does much better job of emphasizing what we actually do.
Joint work with Jeff Negrea. Thread below!
Project page: geospsr.github.io
Paper link: arxiv.org/abs/2503.19136
Link to my student's tweets on this work: x.com/sholalkere/s...
At ICML, we're presenting a paper on uncertainty-aware surface reconstruction!
Compared to previous approaches, we are able to completely remove the need for recursive linear solves for reconstruction and interpolation, using geometric GP machinery.
Check it out!