Advertisement · 728 × 90

Posts by Robert Nowak

Post image

upload.wikimedia.org/wikipedia/en...

2 months ago 3 0 0 0

But, as I’ve heard from others, the AIs often suggest possible connections, related results, or avenues to pursue that hadn’t occured to me. Unfortunately, these are usually dead ends.

3 months ago 1 0 1 0

I’ve found that the pro-version AIs are great for proving known theorems or new theorems that could be considered homework problems, but so far I have had no success using them to solve truly open/novel/challenging math problems.

3 months ago 2 0 1 0
Preview
a man playing a guitar with the words heavy metal intensifies ALT: a man playing a guitar with the words heavy metal intensifies
4 months ago 1 0 1 0

Since I can’t get it out of my head, I wrote up my thoughts on @kevinbaker.bsky.social's critique of AI-automated science and the logical end of processes that can't self-correct.

4 months ago 43 9 3 1

Kevin Baker's essay is probably the best thing I have read in 2025.

4 months ago 1 0 0 0

Yes. Just write your thoughts in a rough and unpolished form, say rough paragraphs that contain terse points you want to make. then let 'er rip

5 months ago 1 0 1 0
Advertisement

Section 7 is a wonderful description of the process they went through.

5 months ago 1 0 1 0

something just isn't fully clicking. if you look at total yards and time of possession, they should have blown them out. well, better anyway to peak later in season, so let's hope that's what happens (like two seasons ago)

6 months ago 1 0 0 0

Packers get the win, but it wasn't pretty.

6 months ago 0 0 1 0

Thanks for participating and presenting your work!

7 months ago 1 0 0 0
Post image

Google promotes box shirts too

7 months ago 1 0 0 0
Post image

Pour into

7 months ago 2 0 0 0
Post image

Announcing the first workshop on Foundations of Language Model Reasoning (FoRLM) at NeurIPS 2025!

📝Soliciting abstracts that advance foundational understanding of reasoning in language models, from theoretical analyses to rigorous empirical studies.

📆 Deadline: Sept 3, 2025

8 months ago 11 3 1 1
Preview
Birdscaping for Wisconsin and the Great Lakes Region by Mariette Nowak If you want the birds to flock to your garden, consider “birdscaping,” advises former director of Milwaukee’s Wehr Nature Center.

Nice article about my mom’s new book shepherdexpress.com/culture/book...

8 months ago 3 0 0 0

“the only way to predict or to control the functioning of such systems is by an intricate system of charms, spells, and incantations”

9 months ago 2 0 0 0
Advertisement

See you there!

9 months ago 2 0 0 0

More likely midges. The truest sign of a healthy ecosystem

11 months ago 1 0 1 0

Looking forward to a great MMLS!

11 months ago 3 0 0 0

This is collaboration with Ziyue Luo, @shroffness and @kevinlauka

1 year ago 0 0 0 0
Preview
GPT-4o as the Gold Standard: A Scalable and General Purpose Approach to Filter Language Model Pretraining Data Large language models require vast amounts of high-quality training data, but effective filtering of web-scale datasets remains a significant challenge. This paper demonstrates that GPT-4o is remarkab...

Jifan’s on the industry job market now, and his expertise in efficient training, distillation, and data curation couldn't be more timely. Feel free to reach out to him at jifan@cs.wisc.edu.
📄 Paper: arxiv.org/abs/2410.02755

1 year ago 1 0 1 0
Post image

SIEVE improves upon existing quality filtering methods in the DataComp-LM challenge, producing better LLM pretraining data that led to improved model performance.
This work is part of Jifan's broader research on efficient ML training, from active learning to label-efficient SFT for LLMs.

1 year ago 0 0 1 0

Why does this matter? High-quality data is the bedrock of LLM training. SIEVE enables filtering trillions of web data for specific domains like medical/legal text with customizable natural language prompts.

1 year ago 0 0 1 0
Post image

SIEVE distills GPT-4's data filtering capabilities into lightweight models at <1% of the cost. Not just minor improvements - we're talking 500x more efficient filtering operations.

1 year ago 0 0 1 0
Post image

🧵 Heard all the buzz around distilling from OpenAI models? Check out @jifanz's latest work SIEVE - showing how strategic distillation can make LLM development radically more cost-effective while matching quality.

1 year ago 4 0 1 1
Post image

Maybe Trump should have read my mom's book: "For the first six weeks, the embryo, whether XX or XY, coasts along in sexual ambiguity." p. 25

1 year ago 2 0 0 0
Advertisement
Preview
Trump's Trans Ban Defines Everyone as Female -- But That's Not the Problem Donald Trump's executive order attempts to get rid of trans rights. LGBTQ+ advocates intend to fight back.

www.rollingstone.com/politics/pol...

1 year ago 0 0 3 0
Post image Post image Post image Post image

Task vectors are akin to punchcards: you feed them to your LLM and it implements specific tasks, without in-context demonstrations. Liu's new paper examines at what scale, where in the network and when during training do they emerge, and how to encourage their emergence.

arxiv.org/pdf/2501.09240

1 year ago 23 3 0 0

Good luck with that

1 year ago 0 0 1 0

p.s. we don't know for sure if I said this or not

1 year ago 1 0 1 0