🎉 25 ELLIS Units have been successfully extended!
Following our five-year reapplication process, we celebrate the sustained excellence these Units bring to European AI research. Congratulations to all! 👏
📖 Get more details: ellis.eu/news/25-elli...
Posts by Visual Inference Lab
📢 Call for Papers: #VMV2026
#GCPR2026 is joined by the International Symposium on Vision, Modeling, and Visualization! We invite high-quality submissions across all areas of visual computing.
We are excited to see your work!
👉 Learn more & submit: www.gcpr-vmv.de/year/2026/vmv
#GCPR2026 is coming to #Siegen! 🎉
Join us for great research at the beautiful Campus Lower Castle of the University of Siegen, perfectly located just a short walk from the central station and many city hotels.
More info: www.gcpr-vmv.de/year/2026/lo...
Got a paper on generative models accepted at CVPR 2026?
Share it with us at the 4th Workshop on Generative Models for Computer Vision!
generative-vision.github.io/workshop-CVP...
You can simply submit your accepted CVPR paper, no need to reformat!
Deadline: April 30 (AoE)
🔥 Can in-context segmentation emerge directly from frozen DINOv3 features?
At #CVPR2026, we present INSID3: Training-Free In-Context Segmentation with DINOv3 — a collaboration between PoliTo, TU Darmstadt and TU Munich.
Check it out: github.com/visinf/INSID3
We are excited to share our call for papers with a submission deadline on May 21st, 2026! We invite submissions of high-quality research papers presenting original contributions in all areas of pattern recognition!
Read more: www.gcpr-vmv.de/year/2026/gc...
#GCPR2026 #VMV2026
📢🎓 We have open PostDoc positions in Computer Vision & ML at @tuda.bsky.social and @hessianai.bsky.social within the Reasonable AI Cluster of Excellence — supervised by @stefanroth.bsky.social, @simoneschaub.bsky.social and many others!
Apply here: www.career.tu-darmstadt.de/tu-darmstadt...
I am one of the potential advisors for this PhD position in our #DAAD funded AI graduate school ELIZA (eliza.school), which connects seven #ELLIS units in Germany. If you are interested in working with one of my colleagues or with me (in computer vision & deep learning), please consider applying.
📢 NeurIPS 2025 Spotlight 📢
Can we embed motion into image representations?
FlowFeat embeds optical flow into pixel-level representations, which results in sharp feature grids, especially for dynamic objects.
Project website: tum-vision.github.io/flowfeat
With Anna Sonnweber and Daniel Cremers.
📢 Join @tuda.bsky.social as a PhD/Postdoc in the new project HAICC - Human–AI Collaboration for Cybersecurity!
Explore how LLM–based AI agents and humans can jointly analyse security data and rethink cybersecurity architectures - supervised by Iryna Gurevych, @stefanroth.bsky.social, and many more!
Attending #Neurips2025? Get your personalized Scholar Inbox conference program now to easily navigate the poster sessions and find what you are looking for:
www.scholar-inbox.com/conference/n...
@neuripsconf.bsky.social is two weeks away!
📢 Stop missing great workshop speakers just because the workshop wasn’t on your radar. Browse them all in one place:
robinhesse.github.io/workshop_spe...
(also available for @euripsconf.bsky.social)
#NeurIPS #EurIPS
📢🎓 We have open PhD positions in Computer Vision & Machine Learning at @tuda.bsky.social and @hessianai.bsky.social within the Reasonable AI Cluster of Excellence — supervised by @stefanroth.bsky.social, @simoneschaub.bsky.social and many others!
www.career.tu-darmstadt.de/tu-darmstadt...
🎉 Today, Simon Kiefhaber will present our ICCV oral paper on how to make optical flow estimators more efficient (faster inference and lower memory usage) with state-of-the-art accuracy:
🌍 visinf.github.io/recover
Talk: Tue 09:30 AM, Kalakaua Ballroom
Poster: Tue 11:45 AM, Exhibit Hall I #76
Interested in 3D DINO features from a single image or unsupervised scene understanding?🦖
Come by our SceneDINO poster at NeuSLAM today 14:15 (Kamehameha II) or Tue, 15:15 (Ex. Hall I 627)!
W/ Jevtić @fwimbauer.bsky.social @olvrhhn.bsky.social Rupprecht, @stefanroth.bsky.social @dcremers.bsky.social
[8/8] 2nd Workshop on Explainable Computer Vision: Quo Vadis?
🌍 excv-workshop.github.io
Sun 8:00 AM — 5:00 PM, Ballroom A
[7/8] Efficient Masked Attention Transformer for Few-Shot Classification and Segmentation
by @dustin-carrion.bsky.social, @stefanroth.bsky.social @simoneschaub.bsky.social
🌍 visinf.github.io/emat
📄 arxiv.org/abs/2507.23642
💻 github.com/visinf/emat
Poster: Sun 4:40 PM - Exhibit Hall II #73
[6/8]
📄 arxiv.org/abs/2509.02545
Talk: Sun 9:30 AM, 306 A
Poster: Sun 10:00 AM - Exhibit Hall II
[6/8] Motion-Refined DINOSAUR for Unsupervised Multi-Object Discovery (Oral at ILR+G Workshop)
by Xinrui Gong*, @olvrhhn.bsky.social *, @christophreich.bsky.social , Krishnakant Singh, @simoneschaub.bsky.social , @dcremers.bsky.social @stefanroth.bsky.social
[5/8] ART: Adaptive Relation Tuning for Generalized Relation Prediction
by Gopika Sudhakaran, Hikaru Shindo, Patrick Schramowski, @simoneschaub.bsky.social, @kerstingaiml.bsky.social, @stefanroth.bsky.social
📄 arxiv.org/abs/2507.23543
Poster: Wed 2:45 PM, Exhibit Hall I #1501
[4/8] Activation Subspaces for Out-of-Distribution Detection
by Baris Zongur, @robinhesse.bsky.social, @stefanroth.bsky.social
📄 arxiv.org/abs/2508.21695
Poster: Tue 11:45 AM, Exhibit Hall I #326
[3/8]
Poster: Tue 3:15 PM, Exhibit Hall I #627
Invited
Invited poster at NeuSLAM workshop: Sun 2:15 PM, 304 A/Exhibit Hall II
[3/8] Feed-Forward SceneDINO for Unsupervised Semantic Scene Completion
by @jev-aleks.bsky.social *, @christophreich.bsky.social *, @fwimbauer.bsky.social , @olvrhhn.bsky.social , Christian Rupprecht, @stefanroth.bsky.social, @dcremers.bsky.social
🌍 visinf.github.io/scenedino/
[2/8] Removing Cost Volumes from Optical Flow Estimators (Oral)
by @skiefhaber.de , @stefanroth.bsky.social @simoneschaub.bsky.social
🌍 visinf.github.io/recover
Talk: Tue 09:30 AM, Kalakaua Ballroom
Poster: Tue 11:45 AM, Exhibit Hall I #76
[1/8] We are presenting four main conference papers, two workshop papers, and a workshop at @iccv.bsky.social 2025 in Hawaii! 🎉🏝
🇨🇳 Meet us at IROS 2025:
- Oral presentation: Thu, 23 Oct, 10:40–10:45 AM, Room 309, at Session ThAT9 – Object Detection, Segmentation and Categorization 5
- Poster session: Starting at 11:10 AM in the same room
For more information:
🌍Project page: vita-epfl.github.io/DFI-OmniSter...
📄 Paper: arxiv.org/pdf/2503.23502
💻 Code: github.com/vita-epfl/DF...
🎥 Video: www.youtube.com/watch?v=CHDQ...
TLDR: Given images captured by two vertically stacked omnidirectional cameras, DFI–OmniStereo integrates a pre-trained monocular depth foundation model into an iterative stereo matching approach. DFI–OmniStereo improves depth estimation accuracy, significantly outperforming the previous SotA.
📢Excited to share our IROS 2025 paper “Boosting Omnidirectional Stereo Matching with a Pre-trained Depth Foundation Model”!
Work by Jannik Endres, @olvrhhn.bsky.social, Charles Cobière, @simoneschaub.bsky.social, @stefanroth.bsky.social and Alexandre Alahi.