Advertisement · 728 × 90

Posts by Dávid Komorowicz

Introduction to Spherical Harmonics for Graphics Programmers

INTRODUCTION TO SPHERICAL HARMONICS FOR GRAPHICS PROGRAMMERS

finally done.

gpfault.net/posts/sph.html

1 week ago 269 82 8 2
Post image

Introducing LoMa, the next generation of feature matcher!

1 week ago 41 4 3 3
https://github.com/sadrasafa/WarpRF

https://github.com/sadrasafa/WarpRF

Sadra released the code for WarpRF 🚀

a training-free uncertainty quantification framework for radiance fields based on multi-view consistency, without training or any changes to the model.

2 weeks ago 5 3 1 0
Post image Post image Post image Post image

Sky2Ground: A Benchmark for Site Modeling under Varying Altitude
Zengyan Wang, Sirshapan Mitra, Rajat Modi, Grace Lim, Yogesh Rawat

tl;dr: Dataset + how to adapt VGGT for satellite + ground imagery. Of course, the method is named SkyNet.
arxiv.org/abs/2603.13740

3 weeks ago 8 1 0 0
Post image Post image Post image Post image

Two #CVPR2026 competitions are live: S23DR 2026 and BuildingWorld 2026!

Task: reconstruct house roof wireframes from point clouds and segmentations:

Total prize fund: $22k
Deadline: end of May 2026
1) huggingface.co/spaces/usm3d...
2) huggingface.co/spaces/Build...
@cvprconference.bsky.social

1 month ago 9 2 2 1
Post image Post image Post image Post image

Global-Aware Edge Prioritization for Pose Graph Initialization

@weitong8591.bsky.social @gtolias.bsky.social Jiri Matas, @danielbarath.bsky.social

tl;dr: another global desc->GNN->MST. Supervision:# triangulated points/pair. +heuristic MST postprocessing. Eval on IMC-PT.
arxiv.org/abs/2602.21963

1 month ago 7 2 0 0

SECOND CALL: SHREC'26 Challenge on 3D Reconstruction

Our dataset features intricate geometries, ideal for benchmarking of high-frequency detail recovery.

All participating will co-author a joint paper submitted to Computers & Graphics.

Track Details shapevision.dcc.uchile.cl/cllull-shrec2026

1 month ago 0 1 0 0

#CVPR2026 reviews are slowly being dispatched over email. Good luck!

2 months ago 8 1 0 1
Screenshot of a paper discussion page titled ‘mHC: Manifold-Constrained Hyper-Connections’. At the top is a card showing the paper title, a ‘View on arXiv’ link, and indicators for 7 posts and 7 researchers. Below are social-style posts referencing the paper: one from ‘NT 5.2 Pyongyang Official™’ linking to arXiv with the caption ‘THE WHALE IS BACK BABYYYY’ and an arXiv preview image, and another from ‘Hacker News’ linking to the same arXiv paper. The interface resembles a research discussion or social feed layout.

Screenshot of a paper discussion page titled ‘mHC: Manifold-Constrained Hyper-Connections’. At the top is a card showing the paper title, a ‘View on arXiv’ link, and indicators for 7 posts and 7 researchers. Below are social-style posts referencing the paper: one from ‘NT 5.2 Pyongyang Official™’ linking to arXiv with the caption ‘THE WHALE IS BACK BABYYYY’ and an arXiv preview image, and another from ‘Hacker News’ linking to the same arXiv paper. The interface resembles a research discussion or social feed layout.

new year and @mariaa.bsky.social and I have some fun new things cooking for the atproto ecosystem...

3 months ago 116 14 10 4
Advertisement

Hey this worked! Import all of your old Twitter posts over to Bluesky (for a few bucks, depending on how irrepressible you were). Now I look like I've been posting on this platform longer than it has existed.

1 year ago 114 24 11 8
Post image

Choosing the right colormap is tricky, too often, they hide subtle details or distort the data. Our new method transforms colormaps to boost local contrast and reveal just noticeable differences, all while keeping the visualization perceptually accurate and accessible.

dl.acm.org/doi/10.1145/...

8 months ago 46 9 1 1
Post image

1/ Can open-data models beat DINOv2? Today we release Franca, a fully open-sourced vision foundation model. Franca with ViT-G backbone matches (and often beats) proprietary models like SigLIPv2, CLIP, DINOv2 on various benchmarks setting a new standard for open-source research.

8 months ago 83 22 2 3
Post image

How can one reconstruct the complete 3D interior of a wood block using only photos of its surfaces? 🪵
At SIGGRAPH'25 (Thursday!), Maria Larsson will present *Mokume*: a dataset of 190 diverse wood samples and a pipeline that solves this inverse texturing challenge. 🧵👇

8 months ago 76 15 2 1
Post image Post image Post image Post image

VLM-Guided Visual Place Recognition for Planet-Scale Geo-Localization

Sania Waheed, Na Min An, Michael Milford , Sarvapali D. Ramchurn, Shoaib Ehsan

tl;dr: in title
arxiv.org/abs/2507.17455

8 months ago 7 2 0 0
Post image Post image Post image Post image

Unposed 3DGS Reconstruction with Probabilistic Procrustes Mapping

Chong Cheng, Zijian Wang, Sicheng Yu, Yu Hu, Nanjie Yao, Hao Wang

tl;dr: submap alignment->point cloud registration->robust Umeyama algorithm->global point cloud and camera trajectory

arxiv.org/abs/2507.18541

8 months ago 5 3 0 0
Post image Post image

New 3D foundation model dropped.

Note: Seems they might have messed up their image matching metrics (seems like acc rather than auc), but should be at least as good as mast3r.

8 months ago 11 2 2 0

Turns out that by default huggingface models run on the CPU...

9 months ago 1 0 0 1

Awesome initiative 🎉
This leaves me wondering though: how come authors attending #EurIPS still have to register for the main #NeurIPS (in the Americas) for their paper to be considered accepted?
You stopped so short of actually allowing ML researchers to fly less!

9 months ago 31 5 5 2
A meme where Anakin and Padme discuss the logics of allowing a NeurIPS event in Europe while forcing authors to also present in the US for publication

A meme where Anakin and Padme discuss the logics of allowing a NeurIPS event in Europe while forcing authors to also present in the US for publication

Sofar it doesn’t look good: neurips.cc/FAQ/AuthorRe...

“At least one author of each accepted paper must register for the main conference. A ‘Virtual Only Pass’ is not sufficient.”

9 months ago 7 2 1 0
Advertisement

WeTransfer just changed their TOS giving themselves permission to train AI on any content you transfer and produce derivative works based on content you transfer that they are allowed to monetize and you are not allowed payment for.

Stop using WeTransfer.

9 months ago 7580 5269 127 462
Post image

The code for our #CVPR2025 paper, PRaDA: Projective Radial Distortion Averaging, is now out!

Turns out distortion calibration from multiview 2D correspondences can be fully decoupled from 3D reconstruction, greatly simplifying the problem

arxiv.org/abs/2504.16499
github.com/DaniilSinits...

9 months ago 12 5 1 0
Video

🦖 We present “Feed-Forward SceneDINO for Unsupervised Semantic Scene Completion”. #ICCV2025
🌍: visinf.github.io/scenedino/
📃: arxiv.org/abs/2507.06230
🤗: huggingface.co/spaces/jev-a...
@jev-aleks.bsky.social @fwimbauer.bsky.social @olvrhhn.bsky.social @stefanroth.bsky.social @dcremers.bsky.social

9 months ago 24 10 1 1
Post image Post image Post image

We just released COLMAP v3.12, which adds long-awaited, end-to-end support for multi-camera rigs and 360° panoramas 👀 COLMAP just got better at handling your robotics, AR/VR, or 360 data - try it yourself and let us know! github.com/colmap/colma... Kudos to Johannes & team for this great work 🚀

9 months ago 22 6 1 0
Post image Post image Post image

Dense Match Summarization for Faster Two-view Estimation

Jonathan Astermark, Anders Heyden, Viktor Larsson
tl;dr: use clustering to reduce RANSAC time when using dense methods like RoMa.
Kudos for eval on WxBS.
P.S. now the same, but for BA?

arxiv.org/abs/2506.028...

9 months ago 12 2 2 1
Video

🤗 I’m excited to share our recent work: TwoSquared: 4D Reconstruction from 2D Image Pairs.
🔥 Our method produces geometry, texture-consistent, and physically plausible 4D reconstructions
📰 Check our project page sangluisme.github.io/TwoSquared/
❤️ @ricmarin.bsky.social @dcremers.bsky.social

11 months ago 9 3 0 1
Video

Can we match vision and language representations without any supervision or paired data?

Surprisingly, yes! 

Our #CVPR2025 paper with @neekans.bsky.social and @dcremers.bsky.social shows that the pairwise distances in both modalities are often enough to find correspondences.

⬇️ 1/4

10 months ago 27 12 1 0
Video

Can you train a model for pose estimation directly on casual videos without supervision?

Turns out you can!

In our #CVPR2025 paper AnyCam, we directly train on YouTube videos and achieve SOTA results by using an uncertainty-based flow loss and monocular priors!

⬇️

11 months ago 25 10 1 1
Advertisement
Video

We also found that this allows the CTM to decide to spend less time thinking on simpler images, thus saving energy. When identifying a gorilla, for example, the CTM’s attention moves from eyes to nose to mouth in a pattern remarkably similar to human visual attention.

11 months ago 18 2 1 0
Post image Post image Post image Post image

High Dynamic Range Novel View Synthesis with Single Exposure

Kaixuan Zhang, Hu Wang, Minxian Li, Mingwu Ren, Mao Ye, Xiatian Zhu

tl;dr:single exposure LDR images in training; LDR image->model+lift->HDR colors; HDR image->LDR image->additional supervision

arxiv.org/abs/2505.01212

11 months ago 1 1 0 0
Video

📢 New paper CVPR 25!
Can meshes capture fuzzy geometry? Volumetric Surfaces uses adaptive textured shells to model hair, fur without the splatting / volume overhead. It’s fast, looks great, and runs in real time even on budget phones.
🔗 autonomousvision.github.io/volsurfs/
📄 arxiv.org/pdf/2409.02482

11 months ago 32 21 1 1