Advertisement Β· 728 Γ— 90

Posts by Luisa Zintgraf

Huge shout-out to my co-first authors @dancalian.bsky.social, @gregfar.bsky.social, & Iurii Kemaev.

And to our amazing collaborators: Matteo Hessel, Jeremy Shar, Junhyuk Oh, AndrΓ‘s GyΓΆrgy, @schaul.bsky.social, @jeffdean.bsky.social, Hado van Hasselt, & Dave Silver.

5 months ago 5 0 1 0

We believe that the DataRater is a promising step towards more automated and principled dataset curation. This could be especially important for filtering and making the best use of massive synthetic datasets in the future.

For a deeper dive, check out arxiv.org/pdf/2505.17895

5 months ago 5 0 1 0
Post image

So what does the DataRater learn? It automatically identifies and down-weights data that aligns with human intuitions of low quality, such as incorrect text encodings, OCR errors, and irrelevant content.

5 months ago 4 0 1 0
Post image

The result? The DataRater is highly effective at filtering data, leading to significant compute efficiency improvements. In our experiments, we observed up to a 46.6% net compute gain while often improving final model performance.

5 months ago 1 1 1 0
Post image

We introduce the DataRater, a meta-learning method that learns to rate the value of each data point for training. Instead of manually specifying filtering rules, we train the DataRater to optimize for a simple goal: improving the training efficiency on a held-out dataset.

5 months ago 3 0 1 0

Foundation models are trained on large datasets, but not all data is created equal. Dataset curation often relies on manual, coarse-grained filtering and hand-crafted rules. This is becoming a major challenge, especially with the rise of synthetic data.

5 months ago 3 0 1 0

Excited to share our new paper, "DataRater: Meta-Learned Dataset Curation"!

We explore a fundamental question: How can we *automatically* learn which data is most valuable for training foundation models?

Paper: arxiv.org/pdf/2505.17895 to appear at @neuripsconf.bsky.social

Thread πŸ‘‡

5 months ago 25 4 1 2
Advertisement

Tagging first author @jakeabeck.bsky.social who just joined bsky! Welcome πŸŽ‰

1 year ago 2 0 0 0
[AUTOML23]  A Tutorial on MetaReinforcement Learning
[AUTOML23] A Tutorial on MetaReinforcement Learning YouTube video by AutoMLConf

πŸ“˜ Journal: nowpublishers.com/article/Deta...
πŸ“ ArXiv: arxiv.org/abs/2301.08028
πŸŽ™οΈ Podcast: www.talkrl.com/episodes/jac...
πŸŽ₯ Talk: youtu.be/XUQ9jLOZqGc

1 year ago 7 1 1 0

πŸŽ‰ Our Meta-RL survey is now published in Foundations and Trends in Machine Learning! A deep dive into how agents can learn to learn πŸ€–πŸ§ 

Huge kudos to Jake Beck & Risto Vuorio for leading the charge, and to co-authors Evan Liu, Zheng Xiong, Chelsea Finn & @shimon8282.bsky.social!

1 year ago 30 3 1 0

Wanna work on Gemini? DeepMind is hiring! πŸš€

1 year ago 4 0 0 0
Preview
Research Scientist, Large Scale Pre-Training Model London, UK

Interested in helping us make Gemini Pro even better?

The Gemini pre-training team is looking for a Research Scientist in London to push the boundaries of LLM scaling: understanding, predicting, and improving. β™ŠοΈπŸš€

Apply here: boards.greenhouse.io/deepmind/job...

1 year ago 8 5 0 1