Advertisement ยท 728 ร— 90

Posts by Bill Psomas @ ICLR 2026

Post image

๐Ÿšจ Efficient Probing (EP) @ #ICLR 2026 ๐Ÿ‡ง๐Ÿ‡ท

Models trained to learn local representations (e.g., MIM) are often undervalued by standard global evaluation.

๐Ÿ‘‰ EP unlocks their potential via attention-based aggregation.

Paper: arxiv.org/abs/2506.10178
Poster ๐Ÿ‘‡

See you in Rio! ๐ŸŒด

3 days ago 2 0 0 0
Preview
Retrieve and Segment: Are a Few Examples Enough to Bridge the Supervision Gap in Open-Vocabulary Segmentation? Open-vocabulary segmentation (OVS) extends the zero-shot recognition capabilities of vision-language models (VLMs) to pixel-level prediction, enabling segmentation of arbitrary categories specified by text prompts. Despite recent progress, OVS lags behind fully supervised approaches due to two challenges: the coarse image-level supervision used to train VLMs and the semantic ambiguity of natural language. We address these limitations by introducing a few-shot setting that augments textual prompts with a support set of pixel-annotated images. Building on this, we propose a retrieval-augmented test-time adapter that learns a lightweight, per-image classifier by fusing textual and visual support features. Unlike prior methods relying on late, hand-crafted fusion, our approach performs learned, per-query fusion, achieving stronger synergy between modalities. The method supports continually expanding support sets, and applies to fine-grained tasks such as personalized segmentation. Experiments show that we significantly narrow the gap between zero-shot and supervised segmentation while preserving open-vocabulary ability.

๐Ÿคœ๐Ÿค› Grateful to @tim-arav.bsky.social, @stojnicv.xyz, Nikos Komodakis, and @gtolias.bsky.social

๐Ÿ“ Paper: arxiv.org/abs/2602.23339
๐Ÿ’ป Code: github.com/TilemahosAra...

#CVPR2026 #ComputerVision #AI #MSCA #MSCAPF #RAG #DL #ML

1 week ago 2 0 0 0
Post image

๐ŸŽ‰ RNS is a Highlight paper at #CVPR 2026 ๐ŸŽ‰

๐Ÿ’ก1.5 years ago, just after finishing my PhD, I wrote my #MSCA PF proposal around a simple idea: can memory extend VLMs for open-vocabulary segmentation?

๐ŸŽ‰ Today, the first paper is a CVPR Highlight.

๐ŸŽฏ From idea โ†’ funded project โ†’ highlight paper.

1 week ago 4 0 1 0
Post image

๐Ÿš€ This week at #ELLIS Winter School: our #CVPR2026 paper Retrieve and Segment (RNS) is being presented by @tim-arav.bsky.social.

๐ŸŽฏ RNS shows how a few pixel-level annotated images can boost zero-shot Open Vocabulary Segmentation.

๐Ÿ“ arxiv.org/pdf/2602.23339

#ComputerVision #FoundationModels #OVS

4 weeks ago 4 0 0 0
Post image

๐Ÿš€ New task: Instance-level Composed Image Retrieval

๐Ÿ”Ž Given a [query image] + [query text], retrieve the particular object after the change โ€” not just any similar object.

๐Ÿ›ข New dataset on HF: i-CIR huggingface.co/datasets/bil...

๐Ÿ“„ Project page: vrg.fel.cvut.cz/icir/

4 weeks ago 2 1 0 0

This is such a good illustration! Too bad I didn't have the idea when I wrote that silly little paper a few years ago.

Anyway, remember folks: torch.manual_seed(3407) is all you need
No other seed had been subject to as much scrutiny as this one
And it's all yours for free!

1 month ago 26 5 2 0
Greeks in AI 2026 Symposium Welcome to the OpenReview homepage for Greeks in AI 2026 Symposium

OpenReview link: openreview.net/group?id=gre...

1 month ago 0 0 0 0

#GreeksInAI #AI #ArtificialIntelligence #MachineLearning #DeepLearning #NLP #ComputerVision #Robotics #MultimodalAI #TrustworthyAI #AIResearch #Innovation #Greece #Athens

1 month ago 1 0 1 0

Important dates
๐Ÿ“ Submission deadline: 25 May 2026
๐Ÿ“ฉ Notification of acceptance: 1 June 2026
๐ŸŽค Dates of the event: 15โ€“17 July 2026

This is a great opportunity to share your work, meet the Greek AI community, and help strengthen the AI ecosystem in Greece and beyond.

1 month ago 2 0 1 0

Submissions may include work submitted or accepted within the last year, and accepted papers will be presented as posters or spotlight talks.

All submissions will be through OpenReview:

1 month ago 1 0 1 0
Advertisement

Topics include, among others:
โ€ข Machine Learning
โ€ข Natural Language Processing
โ€ข Computer Vision
โ€ข Robotics and Autonomous Systems
โ€ข Reinforcement Learning
โ€ข Multimodal AI
โ€ข Trustworthy, Fair, and Explainable AI
โ€ข AI for Science, Healthcare, and Climate
โ€ข AI Applications

1 month ago 4 1 1 0

๐Ÿค– Thrilled to share the Call for Papers for Greeks in AI 2026!

๐Ÿ‡ฌ๐Ÿ‡ท Greeks in AI 2026
๐Ÿ“ Eugenides Foundation, Athens, Greece
๐Ÿ“… 15โ€“17 July 2026

We invite submissions presenting recent advances, ongoing research, and emerging ideas in Artificial Intelligence and Machine Learning.

1 month ago 3 1 1 0
Preview
Retrieve and Segment: Are a Few Examples Enough to Bridge the Supervision Gap in Open-Vocabulary Segmentation? Open-vocabulary segmentation (OVS) extends the zero-shot recognition capabilities of vision-language models (VLMs) to pixel-level prediction, enabling segmentation of arbitrary categories specified by...

7/7 Resources ๐Ÿ“„

Paper: arxiv.org/abs/2602.23339
Code: github.com/TilemahosAra...

Joint work with: @tim-arav.bsky.social, @stojnicv.xyz, Nikos Komodakis, and @gtolias.bsky.social.

We thank @noagarciad.bsky.social, @skamalas.bsky.social, @ekazakos.bsky.social for the feedback.

See you @ #CVPR2026

1 month ago 4 0 0 0
Post image

6/n ๐‘ท๐’†๐’“๐’”๐’๐’๐’‚๐’๐’Š๐’›๐’†๐’… ๐‘บ๐’†๐’ˆ๐’Ž๐’†๐’๐’•๐’‚๐’•๐’Š๐’๐’

๐Ÿฅค ๐‘น๐‘ต๐‘บ can easily be employed for fine-grained tasks like ๐’‘๐’†๐’“๐’”๐’๐’๐’‚๐’๐’Š๐’›๐’†๐’… ๐’”๐’†๐’ˆ๐’Ž๐’†๐’๐’•๐’‚๐’•๐’Š๐’๐’ by simply expanding the support set with a few examples of a specific instance, letting it ๐’”๐’†๐’‘๐’‚๐’“๐’‚๐’•๐’† ๐’•๐’‰๐’‚๐’• ๐’Š๐’๐’”๐’•๐’‚๐’๐’„๐’† ๐’‡๐’“๐’๐’Ž ๐’Š๐’•๐’” ๐’ƒ๐’“๐’๐’‚๐’…๐’†๐’“ ๐’„๐’๐’‚๐’”๐’”.

1 month ago 0 0 1 0
Post image

5/n ๐‘ฉ๐’“๐’Š๐’…๐’ˆ๐’Š๐’๐’ˆ ๐’•๐’‰๐’† ๐‘ฎ๐’‚๐’‘

โšก ๐‘น๐‘ต๐‘บ improves over different kinds of OVS approaches ๐’ƒ๐’š 14.1% ๐’๐’ ๐’‚๐’—๐’†๐’“๐’‚๐’ˆ๐’†, while maintaining open-vocabulary generalization.

1 month ago 0 0 1 0
Post image

4/n ๐‘ซ๐’š๐’๐’‚๐’Ž๐’Š๐’„ ๐‘ญ๐’†๐’˜-๐’”๐’‰๐’๐’• ๐‘บ๐’„๐’†๐’๐’‚๐’“๐’Š๐’๐’”

We investigate multiple ๐’‡๐’†๐’˜-๐’”๐’‰๐’๐’• settings where visual or textual information may be missing for some test classes.

๐ŸŽ‰ We consistently improve respective baselines, making ๐‘น๐‘ต๐‘บ a ๐’‘๐’“๐’‚๐’„๐’•๐’Š๐’„๐’‚๐’ and ๐’๐’‘๐’†๐’-๐’˜๐’๐’“๐’๐’… ๐‘ถ๐‘ฝ๐‘บ method.

1 month ago 0 0 1 0
Post image

3/n ๐‘ฏ๐’๐’˜ it works?

๐Ÿ’พ ๐‘น๐‘ต๐‘บ stores ๐‘ฝ๐‘ณ๐‘ด ๐’‡๐’†๐’‚๐’•๐’–๐’“๐’†๐’” from visual and textual examples in a ๐’Ž๐’†๐’Ž๐’๐’“๐’š-๐’†๐’‡๐’‡๐’Š๐’„๐’Š๐’†๐’๐’• manner.

๐Ÿ–ผ๏ธ At test time, it ๐’“๐’†๐’•๐’“๐’Š๐’†๐’—๐’†๐’” ๐’•๐’†๐’”๐’• ๐’Š๐’Ž๐’‚๐’ˆ๐’† ๐’“๐’†๐’๐’†๐’—๐’‚๐’๐’• ๐’†๐’™๐’‚๐’Ž๐’‘๐’๐’†๐’” to train a linear classifier on both modalities.

1 month ago 0 0 1 0
Post image

2/n Zero-shot open-vocabulary segmentation (OVS) is significantly underperforming fully supervised.

๐ŸŒ‰ ๐‘น๐‘ต๐‘บ ๐’ƒ๐’“๐’Š๐’…๐’ˆ๐’†๐’” ๐’•๐’‰๐’Š๐’” ๐’ˆ๐’‚๐’‘ using a few pixel-level annotated visual examples along with class names.

With a few adaptation steps on each test image, we improve zero-shot ๐’ƒ๐’š up to 34% on average.

1 month ago 0 0 1 0
Advertisement
Video

1/n #CVPR2026 Accepted Paper๐Ÿš€

๐‘จ๐’“๐’† ๐’‚ ๐‘ญ๐’†๐’˜ ๐‘ฌ๐’™๐’‚๐’Ž๐’‘๐’๐’†๐’” ๐‘ฌ๐’๐’๐’–๐’ˆ๐’‰ ๐’•๐’ ๐‘ฉ๐’“๐’Š๐’…๐’ˆ๐’† ๐’•๐’‰๐’† ๐‘บ๐’–๐’‘๐’†๐’“๐’—๐’Š๐’”๐’Š๐’๐’ ๐‘ฎ๐’‚๐’‘ ๐’Š๐’ ๐‘ถ๐’‘๐’†๐’-๐‘ฝ๐’๐’„๐’‚๐’ƒ๐’–๐’๐’‚๐’“๐’š ๐‘บ๐’†๐’ˆ๐’Ž๐’†๐’๐’•๐’‚๐’•๐’Š๐’๐’?

๐‘น๐’†๐’•๐’“๐’Š๐’†๐’—๐’† ๐’‚๐’๐’… ๐‘บ๐’†๐’ˆ๐’Ž๐’†๐’๐’• (๐‘น๐‘ต๐‘บ) answers this question.

Paper/code at the end๐Ÿ‘‡๐Ÿผ

1 month ago 8 2 1 1
Post image Post image Post image Post image

Retrieve and Segment: Are a Few Examples Enough to Bridge the Supervision Gap in Open-Vocabulary Segmentation?

Tilemachos Aravanis @stojnicv.xyz @billpsomas.bsky.social Nikos Komodakis @gtolias.bsky.social

tl;dr: almost yes if use 1-3 images, no if more(fig 6)
arxiv.org/abs/2602.23339
#CVPR2026

1 month ago 7 3 0 0

๐ŸŽ‰๐ŸŽ‰๐ŸŽ‰

1 month ago 1 0 0 0

Excited to share that our paper "Global-Aware Edge Prioritization for Pose Graph Initialization" has been accepted to CVPR 2026! #CVPR2026 See you soon in Denver!๐Ÿฅณ๐Ÿฅณ Code is coming soon๐Ÿšง
โ“How would you do an accurate and efficient pose graph initialization in a global manner? arxiv.org/abs/2602.21963

1 month ago 10 3 1 0
Post image Post image Post image

Global-Aware Edge Prioritization for Pose Graph Initialization

@weitong8591.bsky.social, @gtolias.bsky.social, Jiri Matas, @danielbarath.bsky.social

tl;dr: rank pose graph edges->global consistency->improve SfM

arxiv.org/abs/2602.21963

1 month ago 9 2 1 1
Post image

Sleeping while waiting on an โ€œanywhere in the worldโ€ paper decision release. #CVPR2026

2 months ago 17 1 2 0
Preview
Attention, Please! Revisiting Attentive Probing Through the Lens of Efficiency As fine-tuning becomes impractical at scale, probing is emerging as the preferred evaluation protocol. However, standard linear probing can understate the capability of models whose pre-training optim...

8/8 Resources ๐Ÿ“„

Paper: arxiv.org/abs/2506.10178
Code: github.com/billpsomas/e...

Joint work with: Dionysis Christopoulos,@eirinibaltzi.bsky.social,@ikakogeorgiou.bsky.social, @tim-arav.bsky.social,Nikos Komodakis,Konstantinos Karantzalos,Yannis Avrithis,@gtolias.bsky.social.

See you @ ICLR 2026๐Ÿ‡ง๐Ÿ‡ท

2 months ago 0 0 0 0

7/n Take-home messages ๐Ÿ’ก

EP:
- Plug-and-play.
- Compatible with all pre-training families.
- Unlocks the potential of encoders optimized for local representations.
- Complementary with PEFT.
- Better to have it, than not to have it. ๐Ÿ‘€

2 months ago 0 0 1 0
Post image

6/n EP + PEFT = ๐Ÿ”ฅ

- EP captures information that LoRA alone does not, and vice versa.
- LoRA+EP improves over both pure EP and pure LoRA.

๐Ÿ“Œ Example: a LoRA+EP configuration with 250K params reaches 72%, 4.3% above linear probing (67.7%), while using over 3ร— fewer parameters.

2 months ago 0 0 1 0
Advertisement
Post image Post image

5/n Interpretability ๐Ÿ”

- EP queries specialize in distinct spatial regions.
- Attention maps are complementary.
- Semantic correspondences emerge (e.g. tails, feet).
- Verified quantitatively too.

2 months ago 0 0 1 0
Post image

4/n Designed for local representations๐Ÿงฉ

๐Ÿ“Š Across ImageNet-1K:

- Consistent gains over k-NN and Linear Probing (LP).
- Particularly strong improvements for MIM, VL, and generative.
- Minimal overhead.

2 months ago 0 0 1 0
Post image Post image

3/n Core observation โš™๏ธ

Prior attentive probing uses redundant projections.

๐Ÿ” Introducing Efficient Probing (EP):

๐Ÿ“Œ Multi-query cross-attention.
๐Ÿ”Œ Plug-and-play on top of frozen encoders.
๐Ÿ’ธ Lightweight and parameter-efficient.

2 months ago 0 0 1 0