Advertisement · 728 × 90

Posts by Alexander Terenin

Thank you!

4 weeks ago 0 1 0 0

Thanks! It may look a certain way from outside, but boldness here is a product of conviction and constraints - not one, or the other, but both. Just me doing the best I can to make things happen, however has the chance of working.

1 month ago 3 1 0 0

Thank you!

1 month ago 1 0 0 0
Preview
The Road Less Traveled

Big career news: I'm leaving academia - and moving to the San Francisco Bay Area to explore something new.

I've written a short blog post with a few reflections on the end of this chapter.

If you'd like to catch up, now is the time to reach out!

avt.im/blog/the-roa...

1 month ago 22 1 4 0
Post image

Happy to share a major milestone: after years of development, we are officially launching Version 1.0 of the GeometricKernels library!

To top it off, our accompanying paper has just been published in JMLR (MLOSS)! 🎉

github.com/geometric-ke...

1 month ago 48 12 1 0
Post image

I wrote a long, detailed blog post on the state of AI and machine learning.

The post's purpose is to sharpen my thinking and help ensure I work on the right things over the next few years.

You might find parts of it interesting. Comments are welcome.

avt.im/blog/where-a...

2 months ago 12 0 0 0
Post image

After a delayed flight and missing my first poster, I have officially arrived at NeurIPS!

I’ll present another poster tomorrow at CDE 606 from 11-2. I’ll post more on this soon.

If you’re interested in meeting up, let’s get in touch!

4 months ago 6 0 0 0
Post image

Today, I gave a talk at the INFORMS Job Market Showcase!

If you're interested, here are the slides - link below!

presentations.avt.im/2025-10-26-A...

5 months ago 6 1 0 0
Post image

I am hiring a fully-funded #PhD in #ML to work at the University of Edinburgh on 𝐠𝐞𝐨𝐦𝐞𝐭𝐫𝐢𝐜 𝐥𝐞𝐚𝐫𝐧𝐢𝐧𝐠 and 𝐮𝐧𝐜𝐞𝐫𝐭𝐚𝐢𝐧𝐭𝐲 𝐪𝐮𝐚𝐧𝐭𝐢𝐟𝐢𝐜𝐚𝐭𝐢𝐨𝐧.

Application deadline: 31 Dec '25. Starts May/Sep '26.
Details in the reply.

Pls RT and share with anyone interested!

5 months ago 16 7 1 3
Advertisement
Preview
Bayesian Algorithms for Adversarial Online Learning: from Finite to Infinite Action Spaces We develop a form Thompson sampling for online learning under full feedback - also known as prediction with expert advice - where the learner's prior is defined over the space of an adversary's future...

Check out the work at: arxiv.org/abs/2502.14790

And, again, shoutout to amazing coauthor Jeff Negrea! Working together has been a great pleasure!

Stay tuned for follow-up: we've been working on using this viewpoint to understand other correlated perturbation-based algorithms.

6 months ago 4 0 0 0

So this sums up the work! If you followed along, thanks for the interest!

I think you'd agree that "Bayesian Algorithms for Adversarial Online Learning: from Finite to Infinite Action Spaces" is a much better title than before. The old one was much harder to pronounce.

6 months ago 1 0 1 0

The Bayesian viewpoint proves useful for developing this analysis.

It allows us to guess what a good prior will be, and suggests ways to use probability as a tool to prove the algorithm works.

6 months ago 1 0 1 0

We prove that the Bayesian approach works in this setting too.

To achieve this, we develop a new probabilistic analysis of correlated Gaussian follow-the-perturbed-leader algorithms, of which ours is a special case.

This has been an open challenge in the area.

6 months ago 1 0 1 0

The second one is where X = [0,1]^d and Y is the space of bounded Lipschitz functions.

Here, you can't use a prior with independence across actions. You need to share information between actions.

We do this by using a Gaussian process, with correlations between actions.

6 months ago 1 0 1 0

The first one is the classical discrete setting where standard algorithms such as exponential weights are studied.

You can use a Gaussian prior which is independent across actions.

6 months ago 1 0 1 0

Okay, so we now know what "Bayesian Algorithms for Adversarial Online Learning" are.

What about "from Finite to Infinite Action Spaces"?

This covers the two settings we show the aforementioned results in.

6 months ago 1 0 1 0

This approach appears to not make any sense: the Bayesian model is completely fake.

We're pretending to know a distribution for how the adversary will act in the future.

But, in reality, they can do anything.

And yet... we show that this works!

6 months ago 2 0 1 0

We show that this game secretly has a natural Bayesian strategy - one we show is strong.

What's the strategy?

It's really simple:
- Place a prior distribution of what the adversary will do in the future
- Condition on what the adversary has done
- Sample from the posterior

6 months ago 1 0 1 0
Advertisement

One observation about adversarial online learning is that it appears to have nothing to do with Bayesian learning.

There is a two-player zero-sum game, not a joint probability distribution.

So you can't just solve it by applying Bayes' Rule. Or can you?

6 months ago 1 0 1 0

Okay, so now we understand what "Adversarial Online Learning" is.

We propose "Bayesian Algorithms" for this.

What does that mean? Let's unpack.

6 months ago 1 0 1 0

Online learning is therefore a good model for learning to explore by taking random actions.

In contrast to other approaches to resolving explore-exploit tradeoffs such as upper confidence bounds which produce purely deterministic strategies.

6 months ago 1 0 1 0

So that's our setting. Why's it interesting?

Because many other hard decision problems can be reduced to online learning, including certain forms of reinforcement learning (via decision-estimation coefficients), equilibrium computation (via no-regret dynamics), and others.

6 months ago 1 0 1 0

Specifically, their goal is to minimize regret

R(p,q) = E_{x_t~p_t, y_t~q_t} \sup_{x\in X} \sum_{t=1}^T y_t(x) - \sum_{t=1}^T y_t(x_t).

Meaning, the learner compares the sum of their rewards y_t(x_t) with the sum of y_t(x) for the best possible single non-time-dependent x.

6 months ago 1 0 1 0

The learner's goal is to achieve the highest rewards possible. But at each time, the adversary can choose a different reward function.

So why is this game not impossible?

Because the learner only compares how well they do with the *sum* of the adversary's previous rewards.

6 months ago 1 0 1 0

"Adversarial Online Learning" refers to a two-player zero-sum repeated game between a learner and adversary.

At each time point:
- The learner chooses a distribution of predictions p_t over an action space X.
- The adversary chooses a reward function y_t : X -> R.

6 months ago 1 0 1 0
Advertisement
Preview
Bayesian Algorithms for Adversarial Online Learning: from Finite to Infinite Action Spaces We develop a form Thompson sampling for online learning under full feedback - also known as prediction with expert advice - where the learner's prior is defined over the space of an adversary's future...

First a link: arxiv.org/abs/2502.14790

Now, let's unpack the new title!

Let's start with what we mean by "Adversarial Online Learning".

6 months ago 0 0 1 0
Post image

Paper update: our recent work on Thompson sampling has a shiny new - and I hope much better - name!

This new name does much better job of emphasizing what we actually do.

Joint work with Jeff Negrea. Thread below!

6 months ago 10 3 1 0
Stochastic Poisson Surface Reconstruction with One Solve using Geometric Gaussian Processes

Project page: geospsr.github.io
Paper link: arxiv.org/abs/2503.19136
Link to my student's tweets on this work: x.com/sholalkere/s...

9 months ago 3 0 0 0

At ICML, we're presenting a paper on uncertainty-aware surface reconstruction!

Compared to previous approaches, we are able to completely remove the need for recursive linear solves for reconstruction and interpolation, using geometric GP machinery.

Check it out!

9 months ago 10 3 1 0
Preview
Virtual Seminar Series on Bayesian Decision-making and Uncertainty

Check out all of this season's seminars here: gp-seminar-series.github.io

10 months ago 3 0 0 0