Pretrained ViTs usually come in rigid sizes (S, B, L, H). But your hardware constraints don't
We built a way to make DINO or CLIP fully elastic in <5 mins without any retraining β‘οΈ
Get the exact model size you need, not just what was released
Find Walter at #NeurIPS Poster 4709 | Thu 4:30-7:30 PM
Posts by Yuki Asano
On the occasion of the 1000th citation of our Sinkhorn-Knopp self-supervised representation learning paper, I've written a whole post about the history and the key bits of this method that powers the state-of-the-art SSL vision models.
Read it here :): docs.google.com/document/d/1...
Today, we release Franca, a new vision Foundation Model that matches and often outperforms DINOv2.
The data, the training code and the model weights are open-source.
This is the result of a close and fun collaboration
@valeoai.bsky.social (in France) and @funailab.bsky.social (in Franconia)π
Agreed, very interesting! Future engines that run on information? π€―
Our Lab is now also on bsky! π₯³
ππPaliGemma 2 is our updated and improved PaliGemma release using the Gemma 2 models and providing new pre-trained checkpoints for the full cross product of {224px,448px,896px} resolutions and {3B,10B,28B} model sizes.
1/7
Pls RT
Permanent Assistant Professor (Lecturer) position in Computer Vision @bristoluni.bsky.social [DL 6 Jan 2025]
This is a research+teaching permanent post within MaVi group uob-mavi.github.io in Computer Science. Suitable for strong postdocs or exceptional PhD graduates.
t.co/k7sRRyfx9o
1/2
Today we had a joint workshop between our FunAI Lab, UTN and AIST Japan. 13 talks, 1 cake and lots of Bavarian food really get research discussions going!
Towards more collaborations in AI between π©πͺ & π―π΅.
@hirokatukataoka.bsky.social
Thanks for tagging. In addition have a look at NV-Embed paper: arxiv.org/abs/2405.17428 they do contrastive finetuning after turning on the bidirectional attention mask
Also @phdcomics.bsky.social is on π¦ π. slowly nesting here.
Yay, @xkcd.com is on π¦
Nice π! We love small (M)LLMs :) will training code also be released?
and also perhaps interesting for you: probing text-representations of LLMs for CLIP-like zero-shot classification: arxiv.org/abs/2410.07173
Sam next to his poster; I'm still very impressed he did all this for his MSc thesis! #BMVC2024
exactly. hence the new post-(pre)training term perhaps? post-training seems to be a good generic term for the RLHF/preference tuning etc in NLP allenai.org/papers/tulu-.... so by saying post-pretraining, we could emphasize the fact it's unsupervised
"Post-pretraining", "unsupervised domain adaptation" fits, but I think is used for different tasks
This work was led by Jochem Loedeman in his MSc, and supervised by Maarten Stol, Tengda Han and myself.
π: arxiv.org/abs/2210.06466β¨
Visit BMVC poster 532 at 10am today!
This means we can simply send an adapted RGB image to the server to get a personalised output.
We also show that the gains don't just come from adding a new learnable model, but instead from the interplay between the pretrained one and the PGN.
This CNN (e.g. running on a phone) outputs a softmax over a set of learned tokens. These are then combined and used for the adaptation. This allows efficient learning, but also for moving the signal back into pixel-space via pseudo-inverse.
Also known as reprogramming, works from @phillipisola.bsky.social showed that even adjusting singular pixels allows adapting a model. We take this one step further and make the input-only adaptation signal dependent on the image itself: We introduce a lightweight CNN, the Prompt Generation Network.
LoRA is great but one disadvantage is that if you have 1000s of these adapters and want to serve them in an efficient way, it's very difficult: GPUs are inefficient when you e.g. use one adapter for only one sample in a large batch. The solution is to adapt the model strictly in input-space.
LoRA et al. enable personalised model generation and serving, which is crucial as finetuned models still outperform general ones in many tasks. However, serving a base model with many LoRAs is very inefficient! Now, there's a better way: enter Prompt Generation Networks, presented today #BMVC
Hello world!
Is there any tool to sync twitter and bluesky posting?
My growing list of #computervision researchers on Bsky.
Missed you? Let me know.
go.bsky.app/M7HGC3Y
The thingie that brings over your twitter followers worked jolly well for me. Very cool! I am following another 500 people now thanks to thatβ¦
chromewebstore.google.com/detail/sky-f...