๐จ Efficient Probing (EP) @ #ICLR 2026 ๐ง๐ท
Models trained to learn local representations (e.g., MIM) are often undervalued by standard global evaluation.
๐ EP unlocks their potential via attention-based aggregation.
Paper: arxiv.org/abs/2506.10178
Poster ๐
See you in Rio! ๐ด
Posts by Bill Psomas @ ICLR 2026
๐ค๐ค Grateful to @tim-arav.bsky.social, @stojnicv.xyz, Nikos Komodakis, and @gtolias.bsky.social
๐ Paper: arxiv.org/abs/2602.23339
๐ป Code: github.com/TilemahosAra...
#CVPR2026 #ComputerVision #AI #MSCA #MSCAPF #RAG #DL #ML
๐ RNS is a Highlight paper at #CVPR 2026 ๐
๐ก1.5 years ago, just after finishing my PhD, I wrote my #MSCA PF proposal around a simple idea: can memory extend VLMs for open-vocabulary segmentation?
๐ Today, the first paper is a CVPR Highlight.
๐ฏ From idea โ funded project โ highlight paper.
๐ This week at #ELLIS Winter School: our #CVPR2026 paper Retrieve and Segment (RNS) is being presented by @tim-arav.bsky.social.
๐ฏ RNS shows how a few pixel-level annotated images can boost zero-shot Open Vocabulary Segmentation.
๐ arxiv.org/pdf/2602.23339
#ComputerVision #FoundationModels #OVS
๐ New task: Instance-level Composed Image Retrieval
๐ Given a [query image] + [query text], retrieve the particular object after the change โ not just any similar object.
๐ข New dataset on HF: i-CIR huggingface.co/datasets/bil...
๐ Project page: vrg.fel.cvut.cz/icir/
This is such a good illustration! Too bad I didn't have the idea when I wrote that silly little paper a few years ago.
Anyway, remember folks: torch.manual_seed(3407) is all you need
No other seed had been subject to as much scrutiny as this one
And it's all yours for free!
#GreeksInAI #AI #ArtificialIntelligence #MachineLearning #DeepLearning #NLP #ComputerVision #Robotics #MultimodalAI #TrustworthyAI #AIResearch #Innovation #Greece #Athens
Important dates
๐ Submission deadline: 25 May 2026
๐ฉ Notification of acceptance: 1 June 2026
๐ค Dates of the event: 15โ17 July 2026
This is a great opportunity to share your work, meet the Greek AI community, and help strengthen the AI ecosystem in Greece and beyond.
Submissions may include work submitted or accepted within the last year, and accepted papers will be presented as posters or spotlight talks.
All submissions will be through OpenReview:
Topics include, among others:
โข Machine Learning
โข Natural Language Processing
โข Computer Vision
โข Robotics and Autonomous Systems
โข Reinforcement Learning
โข Multimodal AI
โข Trustworthy, Fair, and Explainable AI
โข AI for Science, Healthcare, and Climate
โข AI Applications
๐ค Thrilled to share the Call for Papers for Greeks in AI 2026!
๐ฌ๐ท Greeks in AI 2026
๐ Eugenides Foundation, Athens, Greece
๐
15โ17 July 2026
We invite submissions presenting recent advances, ongoing research, and emerging ideas in Artificial Intelligence and Machine Learning.
7/7 Resources ๐
Paper: arxiv.org/abs/2602.23339
Code: github.com/TilemahosAra...
Joint work with: @tim-arav.bsky.social, @stojnicv.xyz, Nikos Komodakis, and @gtolias.bsky.social.
We thank @noagarciad.bsky.social, @skamalas.bsky.social, @ekazakos.bsky.social for the feedback.
See you @ #CVPR2026
6/n ๐ท๐๐๐๐๐๐๐๐๐๐๐
๐บ๐๐๐๐๐๐๐๐๐๐๐
๐ฅค ๐น๐ต๐บ can easily be employed for fine-grained tasks like ๐๐๐๐๐๐๐๐๐๐๐๐
๐๐๐๐๐๐๐๐๐๐๐๐ by simply expanding the support set with a few examples of a specific instance, letting it ๐๐๐๐๐๐๐๐ ๐๐๐๐ ๐๐๐๐๐๐๐๐ ๐๐๐๐ ๐๐๐ ๐๐๐๐๐
๐๐ ๐๐๐๐๐.
5/n ๐ฉ๐๐๐
๐๐๐๐ ๐๐๐ ๐ฎ๐๐
โก ๐น๐ต๐บ improves over different kinds of OVS approaches ๐๐ 14.1% ๐๐ ๐๐๐๐๐๐๐, while maintaining open-vocabulary generalization.
4/n ๐ซ๐๐๐๐๐๐ ๐ญ๐๐-๐๐๐๐ ๐บ๐๐๐๐๐๐๐๐
We investigate multiple ๐๐๐-๐๐๐๐ settings where visual or textual information may be missing for some test classes.
๐ We consistently improve respective baselines, making ๐น๐ต๐บ a ๐๐๐๐๐๐๐๐๐ and ๐๐๐๐-๐๐๐๐๐
๐ถ๐ฝ๐บ method.
3/n ๐ฏ๐๐ it works?
๐พ ๐น๐ต๐บ stores ๐ฝ๐ณ๐ด ๐๐๐๐๐๐๐๐ from visual and textual examples in a ๐๐๐๐๐๐-๐๐๐๐๐๐๐๐๐ manner.
๐ผ๏ธ At test time, it ๐๐๐๐๐๐๐๐๐ ๐๐๐๐ ๐๐๐๐๐ ๐๐๐๐๐๐๐๐ ๐๐๐๐๐๐๐๐ to train a linear classifier on both modalities.
2/n Zero-shot open-vocabulary segmentation (OVS) is significantly underperforming fully supervised.
๐ ๐น๐ต๐บ ๐๐๐๐
๐๐๐ ๐๐๐๐ ๐๐๐ using a few pixel-level annotated visual examples along with class names.
With a few adaptation steps on each test image, we improve zero-shot ๐๐ up to 34% on average.
1/n #CVPR2026 Accepted Paper๐
๐จ๐๐ ๐ ๐ญ๐๐ ๐ฌ๐๐๐๐๐๐๐ ๐ฌ๐๐๐๐๐ ๐๐ ๐ฉ๐๐๐
๐๐ ๐๐๐ ๐บ๐๐๐๐๐๐๐๐๐๐ ๐ฎ๐๐ ๐๐ ๐ถ๐๐๐-๐ฝ๐๐๐๐๐๐๐๐๐ ๐บ๐๐๐๐๐๐๐๐๐๐๐?
๐น๐๐๐๐๐๐๐ ๐๐๐
๐บ๐๐๐๐๐๐ (๐น๐ต๐บ) answers this question.
Paper/code at the end๐๐ผ
Retrieve and Segment: Are a Few Examples Enough to Bridge the Supervision Gap in Open-Vocabulary Segmentation?
Tilemachos Aravanis @stojnicv.xyz @billpsomas.bsky.social Nikos Komodakis @gtolias.bsky.social
tl;dr: almost yes if use 1-3 images, no if more(fig 6)
arxiv.org/abs/2602.23339
#CVPR2026
๐๐๐
Excited to share that our paper "Global-Aware Edge Prioritization for Pose Graph Initialization" has been accepted to CVPR 2026! #CVPR2026 See you soon in Denver!๐ฅณ๐ฅณ Code is coming soon๐ง
โHow would you do an accurate and efficient pose graph initialization in a global manner? arxiv.org/abs/2602.21963
Global-Aware Edge Prioritization for Pose Graph Initialization
@weitong8591.bsky.social, @gtolias.bsky.social, Jiri Matas, @danielbarath.bsky.social
tl;dr: rank pose graph edges->global consistency->improve SfM
arxiv.org/abs/2602.21963
Sleeping while waiting on an โanywhere in the worldโ paper decision release. #CVPR2026
8/8 Resources ๐
Paper: arxiv.org/abs/2506.10178
Code: github.com/billpsomas/e...
Joint work with: Dionysis Christopoulos,@eirinibaltzi.bsky.social,@ikakogeorgiou.bsky.social, @tim-arav.bsky.social,Nikos Komodakis,Konstantinos Karantzalos,Yannis Avrithis,@gtolias.bsky.social.
See you @ ICLR 2026๐ง๐ท
7/n Take-home messages ๐ก
EP:
- Plug-and-play.
- Compatible with all pre-training families.
- Unlocks the potential of encoders optimized for local representations.
- Complementary with PEFT.
- Better to have it, than not to have it. ๐
6/n EP + PEFT = ๐ฅ
- EP captures information that LoRA alone does not, and vice versa.
- LoRA+EP improves over both pure EP and pure LoRA.
๐ Example: a LoRA+EP configuration with 250K params reaches 72%, 4.3% above linear probing (67.7%), while using over 3ร fewer parameters.
5/n Interpretability ๐
- EP queries specialize in distinct spatial regions.
- Attention maps are complementary.
- Semantic correspondences emerge (e.g. tails, feet).
- Verified quantitatively too.
4/n Designed for local representations๐งฉ
๐ Across ImageNet-1K:
- Consistent gains over k-NN and Linear Probing (LP).
- Particularly strong improvements for MIM, VL, and generative.
- Minimal overhead.
3/n Core observation โ๏ธ
Prior attentive probing uses redundant projections.
๐ Introducing Efficient Probing (EP):
๐ Multi-query cross-attention.
๐ Plug-and-play on top of frozen encoders.
๐ธ Lightweight and parameter-efficient.