Study in Nature: “Across 30 out of 32 evaluation axes from the specialist physician perspective & 25 out of 26 evaluation axes from the patient-actor perspective, AMIE [Google Medical LLM] was rated superior to PCPs [primary care docs] while being non-inferior on the rest.”
(& AIME is an older LLM)
Posts by Kuzman Ganchev
Gemma 3 explained: Longer context, image support, and a new 1B model. → goo.gle/4lV8iaw
Other key enhancements:
🔸 Best model that fits in a single consumer GPU or TPU host
🔸 KV-cache memory reduction with 5-to-1 interleaved attention
🔸 And more!
Read the blog for the full details on Gemma 3.
There's a link to a really nice interactive viewer for a sample of the data (will only make sense after you read the post). There's some examples that I would have expected (where something is implied but not directly stated) but also a surprising number of kind of topical things.
Want to get started using PaliGemma 2?
🎤 developers.googleblog.com/en/introduci...
🤗 huggingface.co/blog/paligem...
💾 kaggle.com/models/googl...
🔧 github.com/google-resea...
7/7
Wanted to share that Varun Godbole recently released a prompting playbook. The title says prompt tuning, but this is text prompts, not soft prompts.
github.com/varungodbole...
I’m pretty excited about this one!
ALTA is A Language for Transformer Analysis.
Because ALTA programs can be compiled to transformer weights, it provides constructive proofs of transformer expressivity. It also offers new analytic tools for *learnability*.
arxiv.org/abs/2410.18077