Advertisement · 728 × 90

Posts by bilal

it's Space Quebec, because they speak a language with the same phonology as French which is incomprehensible to Francophones

11 months ago 221 26 15 1
Preview
Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction We present Visual AutoRegressive modeling (VAR), a new generation paradigm that redefines the autoregressive learning on images as coarse-to-fine "next-scale prediction" or "next-resolution prediction...

there’s been some interesting work lately on multiscale autoregressive image modeling arxiv.org/abs/2404.029...

1 year ago 6 0 1 0

city of stairs and the tainted cup both very good 👍

1 year ago 1 0 0 0

once again coming crawling back to AdamW after every paper published after 2015 has failed me again

1 year ago 67 4 6 0
1 year ago 2 0 0 0
Post image

i think they’re just starting to realize what they unleashed w the tweet

1 year ago 1 0 0 0
Post image

from the other app xd

1 year ago 1 1 0 0

Love to log on to the Horrors app to catch up on today's Horrors

1 year ago 20 3 1 0
Advertisement

we’re doing ai exorcisms in 2025 huh

1 year ago 8 0 0 0

like one of the big things about 404 media, Brian Merchant, Paris Marx, Ed Zitron, etc., is that they neither know nor care how the subject of their criticism actually works

1 year ago 180 15 12 10

I wish academic ML was a bit more skeptical of papers and less skeptical of industry. I get that it sucks to not have visibility on details, but it doesn’t invalidate the results. On the flip side, there are too many papers whose message are parroted despite sketchy experiments.

1 year ago 9 2 3 0

we’re going shopping

1 year ago 1 0 0 0

an LLM that uses streetview to pre-drive down the route and assemble comments like "at the big red barn, turn left" "when you get to the sorta squiggly road, take the exit" like a farmer would

1 year ago 6 3 0 0

if you squint hard enough everything in ml is either a special case of the KL div or newton’s method

1 year ago 3 0 0 0

a lot of machine learning research is about discovering which parts of mathematics are actually L2 regularization and which parts of mathematics are actually Adam

1 year ago 76 5 6 1
1 year ago 152 9 2 0
Advertisement

justine tunney the libc mutex micro optimizations person??

1 year ago 23 0 1 1

This guy needs to read Manufacturing Consent! You’re not supposed to do this yourself you gotta hire editors who already agree with you, this is amateur hour shit…

1 year ago 6 2 3 0
1 year ago 639 82 3 0

pov: post training researchers learning what pretraining researchers do while waiting for the model to train

1 year ago 4 0 0 0

accidentally typed rm -fr and i’m using that now

1 year ago 612 44 46 13

congrats!!

1 year ago 1 0 0 0

thanks for cleaning it up

1 year ago 1 0 0 0
Advertisement
Post image

ai generated slavoj zizek voice on slop video of some bizarre rural chinese cooking

1 year ago 58 2 1 1

incredible new forms of postings emerging

1 year ago 3 0 1 0

interesting is there anywhere i can read more about this

1 year ago 1 0 1 0

its all approximating numbers w other numbers all the way down. everything else is an implementation detail! 😛

1 year ago 3 0 1 0

if your values do matter replace them w values similar to them aka ones (parameter sharing / shared kv cache / factorizing a large matrix into two small ones / lora / adafactor)

1 year ago 4 0 1 0

i love how every efficiency advance in machine learning is approximating [expensive operation] by either ones (just pass it straight through) or zeros (doesn’t matter, just don’t compute it/sparsity)

1 year ago 20 1 2 0