Advertisement ยท 728 ร— 90

Posts by Jon Barron

The trick is using the AI to write you smaller versions of the experiments that you can iterate on faster

1 week ago 1 0 1 0

I would just ask them about their methodology and their interpretation in the normal human way that we currently do. It's very easy to tell the difference between a PhD student who is using an LLM to do science and someone who is just along for the ride as the LLM does science for them.

1 week ago 1 0 0 0

They probably could if I was on their committee, which happens sometimes.

2 weeks ago 0 0 1 0

yeah I guess definitionally if someone makes new knowledge, you could say that they are learning the knowledge that they make by virtue of it being in their head? But that's not usually how people talk about new discoveries or findings being made.

1 month ago 0 0 0 0

ah I guess my program fast-tracked me

1 month ago 0 0 0 0

My understanding was that the goal of doing a PhD was to make new knowledge

1 month ago 3 0 4 0

yeah that was an odd exchange, I didn't expect that concern. I guess I don't really think of a PhD has being primarily about the student learning stuff, but instead about the student accomplishing stuff (or learning how to accomplish stuff).

1 month ago 1 0 2 0
Advertisement
Post image Post image Post image Post image

Radiance Meshes for Volumetric Reconstruction

Alexander Mai, Trevor Hedstrom, @grgkopanas.bsky.social, Janne Kontkanen, Falko Kuester, @jonbarron.bsky.social

tl;dr: Delaunay tetrahedralization->constant density and linear color radian->radiance mesh->radiance field

arxiv.org/abs/2512.04076

4 months ago 3 1 0 0

This, combined with most fields outside of computer science being overly concerned with maintaining cultural and social solidarity (especially against encroaching technology brothers) seems like the most likely explanation.

5 months ago 1 0 0 0
Post image

Is basic image understanding solved in todayโ€™s SOTA VLMs? Not quite.

We present VisualOverload, a VQA benchmark testing simple vision skills (like counting & OCR) in dense scenes. Even the best model (o3) only scores 19.8% on our hardest split.

7 months ago 17 5 2 0
Video

Hereโ€™s what Iโ€™ve been working on for the past year. This is SkyTour, a 3D exterior tour utilizing Gaussian Splat. The UX is in the modeling of the โ€œflight path.โ€ I led the prototyping team that built the first POC. I was the sole designer and researcher on the project, one of the 1st inventors.

9 months ago 66 5 5 0

Ah cool, then why is that last bit true?

9 months ago 0 0 0 0

I don't see how the last sentence follows logically from the two prior sentences.

9 months ago 0 0 1 0

Be sure to do a dedication where you thank a ton of people, it's kind plus it feels good.

Besides that I'd just do a staple job of your papers. Doing new stuff in a thesis is usually a mistake, unless you later submit it as a paper or post it online somewhere. Nobody reads past the dedication.

9 months ago 3 0 1 0

This thread rules

9 months ago 3 0 0 0
SpAItial AI: Building Spatial Foundation Models
SpAItial AI: Building Spatial Foundation Models YouTube video by SpAItial AI

๐Ÿš€๐Ÿš€๐Ÿš€Announcing our $13M funding round to build the next generation of AI: ๐’๐ฉ๐š๐ญ๐ข๐š๐ฅ ๐…๐จ๐ฎ๐ง๐๐š๐ญ๐ข๐จ๐ง ๐Œ๐จ๐๐ž๐ฅ๐ฌ that can generate entire 3D environments anchored in space & time. ๐Ÿš€๐Ÿš€๐Ÿš€

Interested? Join our world-class team:
๐ŸŒ spaitial.ai

youtu.be/FiGX82RUz8U

10 months ago 52 9 4 0

๐Ÿ“บ Now available: Watch the recording of Aaron Hertzmann's talk, "Can Computers Create Art?" www.youtube.com/watch?v=40CB...
@uoftartsci.bsky.social

11 months ago 10 4 0 0
Advertisement
Radiance Fields and the Future of Generative Media
Radiance Fields and the Future of Generative Media YouTube video by Jon Barron

Here's a recording of my 3DV keynote from a couple weeks ago. If you're already familiar with my research, I recommend skipping to ~22 minutes in where I get to the fun stuff (whether or not 3D has been bitter-lesson'ed by video generation models)

www.youtube.com/watch?v=hFlF...

11 months ago 60 12 2 1
Login โ€ข Instagram Welcome back to Instagram. Sign in to check out what your friends, family & interests have been capturing & sharing around the world.

www.instagram.com/mrtoledano/ for anyone else who wanted to see more of this artist's work, really cool stuff!

11 months ago 1 0 0 0

yeah those fisher kernel models were surprisingly gnarly towards the end of their run.

1 year ago 1 0 0 0

yep absolutely. Super hard to do, but absolutely the best approach if it works.

1 year ago 4 0 0 0

If you want you can see the models that AlexNet beat in the 2012 imagenet competition, they were quite huge, here's one: www.image-net.org/static_files.... But I think the better though experiment is to imagine how large a shallow model would have to be to match AlexNet's capacity (very very huge)

1 year ago 2 0 1 0
Post image

One pattern I like (used in DreamFusion and CAT3D) is to "go slow to go fast" --- generate something small and slow to harness all that AI goodness, and then bake that 3D generation into something that renders fast. Moving along this speed/size continuum is a powerful tool.

1 year ago 7 0 1 0
Post image

It makes sense that radiance fields trended towards speed --- real-time performance is paramount in 3D graphics. But what we've seen in AI suggests that magical things can happen if you forgo speed and embrace compression. What else is in that lower left corner of this graph?

1 year ago 4 0 2 0
Post image

And this gets a bit hand-wavy, but NLP also started with shallow+fast+big n-gram models, then moved to parse trees etc, and then on to transformers. And yes, I know, transformers aren't actually small, but they are insanely compressed! "Compression is intelligence", as they say.

1 year ago 3 0 1 0
Advertisement
Post image

In fact, it's the *opposite* of what we saw in object recognition. There we started with shallow+fast+big models like mixtures of Gaussians on color, then moved to more compact and hierarchical models using trees and features, and finally to highly compressed CNNs and VITs.

1 year ago 4 0 2 0
Post image

Let's plot the trajectory of these three generations, with speed on the x-axis and model size on the y-axis. Over time, we've been steadily moving to bigger and faster models, up and to the right. This is sensible, but it's not the trend that other AI fields have been on...

1 year ago 5 0 1 0
Post image

Generation three swapped out those voxel grids for a bag of particles, with 3DGS getting the most adoption (shout out to 2021's pulsar though). These models are larger than grids, and can be tricky to optimize, but the upside for rendering speed is so huge that it's worth it.

1 year ago 7 0 1 0
Post image

The second generation was all about swapping out MLPs for a giant voxel grid of some kind, usually with some hierarchy/aliasing (NGP) or low-rank (TensoRF) trick for dealing with OOMs. These grids are much bigger than MLPs, but they're easy to train and fast to render.

1 year ago 4 0 1 0