Advertisement ยท 728 ร— 90

Posts by Nicolas Beltran-Velez

Post image

๐ŸŽ“ Hats off to the 2025 IICD graduates: Yining Ma Junze Huang Yichi Yang Ruilin Dai Boan Zhu Cameron Park @jlfan.bsky.social & Achille Nazaret!
Wishing you all the best in your next chapter โ€” weโ€™re proud of you! ๐Ÿ’™ #Columbia2025
@bleilab.bsky.social @khanhndinh.bsky.social @elhamazizi.bsky.social

11 months ago 5 1 0 0
Preview
Why Knowledge Distillation Works in Generative Models: A Minimal Working Explanation Knowledge distillation (KD) is a core component in the training and deployment of modern generative models, particularly large language models (LLMs). While its empirical benefits are well documented-...

this is probably not the complete picture of KD, but i can definitely sleep better after writing down and confirming this minimal working explanation.

arXiv: arxiv.org/abs/2505.13111

(3/4)

11 months ago 4 2 2 0
Post image

I received a review like this five years ago. Itโ€™s probably the right time now to share it with everyone who wrote or got random discouraging reviews from ICML/ACL.

1 year ago 65 5 1 3
Post image

First 11 chapters of RLHF Book have v0 draft done. Should be useful now.

Next:
* Crafting more blog content into future topics,
* DPO+ chapter,
* Meeting with publishers to get wheels turning on physical copies,
* Cleaning & cohesiveness
rlhfbook.com

1 year ago 48 9 0 0
Post image

๐Ÿ”ฅ Benchmark Alert! MotifBench sets a new standard for evaluating protein design methods in motif scaffolding.
Why does this matter? Reproducibility & fair comparison have been lackingโ€”until now.
Paper: arxiv.org/abs/2502.12479 | Repo: github.com/blt2114/Moti...
A thread โฌ‡๏ธ

1 year ago 41 17 1 5
Post image

The HuggingFace/Nanotron team just shipped an entire pretraining textbook in interactive format. huggingface.co/spaces/nanot...

Itโ€™s not just a great pedagogic support, but many unprecedented data and experiments presented for the first time in a systematic way.

1 year ago 39 9 0 0
Post image

I just wanted to see what it looked like ๐Ÿ˜ญ

1 year ago 3 0 0 0

Good God, please. I just want some gradients that don't vanish ๐Ÿ˜ญ

1 year ago 4 0 1 0

I was hoping that recent events would lead to a mass exodus from X. Many have left, but most of the ML and LLM people have not.

I have lost a lot of respect for the ML community.

1 year ago 72 4 9 2
Video

Now that bluesky has gifs (it didn't work?), I can share (again) my educational notebook on discrete flow matching (by Itai Gat et al.). Also please check the original article and official implementation by Meta!

๐Ÿ github.com/gle-bellier/...
๐Ÿ github.com/facebookrese...
๐Ÿ“„ arxiv.org/abs/2407.15595

1 year ago 15 2 1 0
Advertisement
Preview
SDE Matching: Scalable and Simulation-Free Training of Latent Stochastic Differential Equations The Latent Stochastic Differential Equation (SDE) is a powerful tool for time series and sequence modeling. However, training Latent SDEs typically relies on adjoint sensitivity methods, which depend ...

Really excited about this! We note a connection between diffusion/flow models and neural/latent SDEs. We show how to use this for simulation-free learning of fully flexible SDEs. We refer to this as SDE Matching and show speed improvements of several orders of magnitude.

arxiv.org/abs/2502.02472

1 year ago 50 10 0 0
This is a scatterplot with the following key features:

Axes:
The x-axis represents "Interest in AI," with values ranging approximately from -2 to 2.
The y-axis represents "Willingness to Tolerate Closed, Autocratic Systems," also ranging from about -2 to 2.
Data Points:
Black dots dominate the plot, distributed across all four quadrants, indicating diverse positions on both variables.
A few red dots labeled "my peeps" are clustered in the bottom-right quadrant, signifying high interest in AI but low tolerance for closed, autocratic systems.
Blue Lines:
The plot includes horizontal and vertical blue lines at zero, dividing it into four quadrants for visual reference.
This visualization highlights a subset of individuals ("my peeps") who stand out from the majority based on their distinct combination of interest and values.

This is a scatterplot with the following key features: Axes: The x-axis represents "Interest in AI," with values ranging approximately from -2 to 2. The y-axis represents "Willingness to Tolerate Closed, Autocratic Systems," also ranging from about -2 to 2. Data Points: Black dots dominate the plot, distributed across all four quadrants, indicating diverse positions on both variables. A few red dots labeled "my peeps" are clustered in the bottom-right quadrant, signifying high interest in AI but low tolerance for closed, autocratic systems. Blue Lines: The plot includes horizontal and vertical blue lines at zero, dividing it into four quadrants for visual reference. This visualization highlights a subset of individuals ("my peeps") who stand out from the majority based on their distinct combination of interest and values.

I have a sinking feeling that by 2029 I'm going to be faking a British accent so no one will think I was one of the *Americans* working on AI during the regime.

1 year ago 110 10 9 0

NGL, it's kind of surprising that more people haven't migrated here, especially given what Musk has been doing these days. I don't get it.

1 year ago 1 0 0 0
Post image

Since everyone wants to learn RL for language models now post DeepSeek, reminder that I've been working on this book quietly in the background for months.

Policy gradient chapter is coming together. Plugging away at the book every day now.

rlhfbook.com/c/11-policy-...

1 year ago 155 19 2 1

Please stop anthropomorphizing language models, it makes them feel really bad

1 year ago 70 2 3 0
Preview
From the fednews community on Reddit Explore this post and more from the fednews community

This comments section is the first time I've felt even a shred of hope in eight days.

1 year ago 20392 3848 568 645
Preview
Democracy 2025 | The united legal frontline in the fight for our democracy Democracy 2025 is the strategic hub to protect people and their rights should the Trump-Vance administration seek to unlawfully strip away freedoms and prosperity.

democracy2025.org/response-center keeps track of it.

1 year ago 5 1 1 0

Nazi salutes and speaking at neo-Nazi rallies seems bad. There's history that we should learn from.

1 year ago 115 5 3 0

Something I really like about NLP research is that it makes everything super intuitive. This week I have been thinking about variational inference in NLP and a lot of the things that seemed to require mathematical intuition just become trivial when thinking about language. So cool:)

1 year ago 2 0 0 0
Advertisement

Tests!! :)

1 year ago 3 0 0 0

But the memory needed for the value function kills the ones that don't have good GPUs ๐Ÿ˜ญ

1 year ago 0 0 0 0
Post image Post image

New randomized, controlled trial by the World Bank of students using GPT-4 as a tutor in Nigeria. Six weeks of after-school AI tutoring = 2 years of typical learning gains, outperforming 80% of other educational interventions.

And it helped all students, especially girls who were initially behind.

1 year ago 354 88 15 27

I mostly use copilot for writing code (as auto complete), gpt4-o for boiler plate, and o1 for serious debugging or boilerplate with some complexity or a lot of requirements. I also use o1 for quick but slightly involved experiments but not as often.

1 year ago 5 0 0 0

I use chatgpt over Google for a lot of things because it is really good at fuzzy queries + data aggregation from many sources. I feel that as long as you double check results it is much faster and convenient.

1 year ago 4 0 0 0

Does anyone have any good resources to learn about quantization? Any essential papers to read and resources about how to use/quantize models in practice are greatly appreciated!

1 year ago 3 0 0 0

That seems very reasonable no? I assumed the public set was used by most (almost all?) algorithms that have been benchmarked against the task. Isn't that the case? (legitimate question)

1 year ago 4 0 0 0

1-> 2 -> 3 -> 3.5 -> 4 -> 4o -> o1 -> o3

I guess we need AGI just to figure out how to name things

1 year ago 71 6 7 0
Advertisement

If you are into ML theory (RL or not) with a proven track record, and you are interested in an industry research position, PM me. Feel free to spread the word.

1 year ago 74 31 2 0

๐Ÿงต Excited to share #Echidna, a Bayesian framework for quantifying the impact of gene dosage on phenotypic plasticity: tinyurl.com/296kf7hf!
With @elhamazizi.bsky.social and @mingxz.bsky.social, we integrate scRNA-seq & WGS to uncover how CNAs drive tumor evolution and transcriptional variability.

1 year ago 15 6 2 2