๐ฅ Can in-context segmentation emerge directly from frozen DINOv3 features?
At #CVPR2026, we present INSID3: Training-Free In-Context Segmentation with DINOv3 โ a collaboration between PoliTo, TU Darmstadt and TU Munich.
Check it out: github.com/visinf/INSID3
Posts by Oliver Hahn
๐ข NeurIPS 2025 Spotlight ๐ข
Can we embed motion into image representations?
FlowFeat embeds optical flow into pixel-level representations, which results in sharp feature grids, especially for dynamic objects.
Project website: tum-vision.github.io/flowfeat
With Anna Sonnweber and Daniel Cremers.
@neuripsconf.bsky.social is two weeks away!
๐ข Stop missing great workshop speakers just because the workshop wasnโt on your radar. Browse them all in one place:
robinhesse.github.io/workshop_spe...
(also available for @euripsconf.bsky.social)
#NeurIPS #EurIPS
๐ข๐ We have open PhD positions in Computer Vision & Machine Learning at @tuda.bsky.social and @hessianai.bsky.social within the Reasonable AI Cluster of Excellence โ supervised by @stefanroth.bsky.social, @simoneschaub.bsky.social and many others!
www.career.tu-darmstadt.de/tu-darmstadt...
Interested in 3D DINO features from a single image or unsupervised scene understanding?๐ฆ
Come by our SceneDINO poster at NeuSLAM today 14:15 (Kamehameha II) or Tue, 15:15 (Ex. Hall I 627)!
W/ Jevtiฤ @fwimbauer.bsky.social @olvrhhn.bsky.social Rupprecht, @stefanroth.bsky.social @dcremers.bsky.social
๐ขExcited to share our IROS 2025 paper โBoosting Omnidirectional Stereo Matching with a Pre-trained Depth Foundation Modelโ!
Work by Jannik Endres, @olvrhhn.bsky.social, Charles Cobiรจre, @simoneschaub.bsky.social, @stefanroth.bsky.social and Alexandre Alahi.
๐ Looking for a PhD position in computer vision? Apply to the European Laboratory for Learning & Intelligent Systems (ELLIS) and work with @stefanroth.bsky.social & @simoneschaub.bsky.social! Join the info session on Oct 1.
@ellis.eu @tuda.bsky.social
ellis.eu/news/ellis-p...
Some impressions from our VISINF summer retreat at Lizumer Hรผtte in the Tirol Alps โ including a hike up Geier Mountain and new research ideas at 2,857 m! ๐ฆ๐น๐๏ธ
๐๐ฒ๐ฒ๐ฑ-๐๐ผ๐ฟ๐๐ฎ๐ฟ๐ฑ ๐ฆ๐ฐ๐ฒ๐ป๐ฒ๐๐๐ก๐ข ๐ณ๐ผ๐ฟ ๐จ๐ป๐๐๐ฝ๐ฒ๐ฟ๐๐ถ๐๐ฒ๐ฑ ๐ฆ๐ฒ๐บ๐ฎ๐ป๐๐ถ๐ฐ ๐ฆ๐ฐ๐ฒ๐ป๐ฒ ๐๐ผ๐บ๐ฝ๐น๐ฒ๐๐ถ๐ผ๐ป
Aleksandar Jevtiฤ, Christoph Reich, Felix Wimbauer ... Daniel Cremers
arxiv.org/abs/2507.06230
Trending on www.scholar-inbox.com
๐ฆ We present โFeed-Forward SceneDINO for Unsupervised Semantic Scene Completionโ. #ICCV2025
๐: visinf.github.io/scenedino/
๐: arxiv.org/abs/2507.06230
๐ค: huggingface.co/spaces/jev-a...
@jev-aleks.bsky.social @fwimbauer.bsky.social @olvrhhn.bsky.social @stefanroth.bsky.social @dcremers.bsky.social
Congrats! ๐
Got a strong XAI paper rejected from ICCV? Submit it to our ICCV eXCV Workshop todayโwe welcome high-quality work!
๐๏ธ Submissions open until June 26 AoE.
๐ Got accepted to ICCV? Congrats! Consider our non-proceedings track.
#ICCV2025 @iccv.bsky.social
Scene-Centric Unsupervised Panoptic Segmentation
by @olvrhhn.bsky.social , @christophreich.bsky.social , @neekans.bsky.social , @dcremers.bsky.social, Christian Rupprecht, and @stefanroth.bsky.social
Sunday, 8:30 AM, ExHall D, Poster 330
Project Page: visinf.github.io/cups
Can you train a model for pose estimation directly on casual videos without supervision?
Turns out you can!
In our #CVPR2025 paper AnyCam, we directly train on YouTube videos and achieve SOTA results by using an uncertainty-based flow loss and monocular priors!
โฌ๏ธ
Happy to be recognized as an Outstanding Reviewer at #CVPR2025 ๐
๐Excited to announce our CVPR 2025 paper: Unbiasing through Textual Descriptions!
We release new descriptions for 1.9M(!) videos and object-debiased splits for 12 datasets!
๐Project: utd-project.github.io
by @ninashv.bsky.social et al ๐งต๐
@cvprconference.bsky.social
Check out our latest recent #CVPR2025 paper AnyCam, a fast method for pose estimation in casual videos!
1๏ธโฃ Can be directly trained on casual videos without the need for 3D annotation.
2๏ธโฃ Based around a feed-forward transformer and light-weight refinement.
Code and more info: โฉ fwmb.github.io/anycam/
๐ Never miss a beat in science again!
๐ฌ Scholar Inbox is your personal assistant for staying up to date with your literature. It includes: visual summaries, collections, search and a conference planner.
Check out our white paper: arxiv.org/abs/2504.08385
#OpenScience #AI #RecommenderSystems
Excited to share our #CVPR2025 highlight on unsupervised panoptic segmentation!
We are thrilled to have 12 papers accepted to #CVPR2025. Thanks to all our students and collaborators for this great achievement!
For more details check out cvg.cit.tum.de
๐๏ธโท๏ธ Looking back on a fantastic week full of talks, research discussions, and skiing in the Austrian mountains!
Our work, "Boosting Unsupervised Semantic Segmentation with Principal Mask Proposals" is accepted at TMLR! ๐
visinf.github.io/primaps/
PriMaPs generate masks from self-supervised features, enabling to boost unsupervised semantic segmentation via stochastic EM.