Advertisement · 728 × 90

Posts by Elixir of Progress

You cannot use economics to solve an epistemology or education problem, except perhaps to direct resources towards giving people better education and epistemology.

2 hours ago 3 0 0 0

You can't actually transport DC electricity very easily.

3 hours ago 1 0 0 0

Oh no please, please. I hope Terrance Howard's math thing and catching on.

4 hours ago 2 0 0 0

Trying to explain St Augustine to the pope, the former head of the Augustinian order, who wrote his doctoral thesis on Augustine, on his way back from celebrating mass at the Basilica of St Augustine in Annaba, Algeria, overlooking the site where Augustine lived is peak Adult Catholic Convert.

8 hours ago 4892 1560 98 99

I wish she'd actually apologize for the time she threw NB folks under the bus and then said, "I don't have to apologize, I used to be NB." That was not very nice of her.

4 hours ago 6 0 1 0

Based.

4 hours ago 4 0 1 0

The ugly bit is the conversion mechanism: scientific judgment gets deskilled into verification piecework.

Not so much "the AI replaces PhDs", but "PhDs become the unpaid/underpaid checksum layer for the machine that will be marketed as replacing them". Very normal civilization.

4 hours ago 15 3 1 0
Advertisement
Preview
How Silicon Valley Is Turning Scientists Into Exploited Gig Workers Tech elites are enriching themselves by plundering STEM institutions—and offering researchers scraps.

You know what's neat about this? It's basically labor arbitrage built out of institutional sabotage. Many such cases, I know.

4 hours ago 34 13 1 0

I'd like to see your very rugged solar panels that can survive having American land barges drive over them.

5 hours ago 2 0 2 0

That's surprising to me. What makes France so cheap?

5 hours ago 0 0 1 0

Are us construction costs really that much higher than France's?

5 hours ago 4 0 1 0

In what context would construction be *cheaper* than over a parking lot?

5 hours ago 3 0 1 0

Gemma 26B A4B (however Ollama quant'd it for now, but my hardware can accommodate it in fp16 if I wanna take the performance hit).

9 hours ago 2 0 1 0

It's the 26B A4B specifically.

Using an MoE model is a special discipline. You need to be *super* careful about context locking you, but Gemma4's router is doing things I don't understand to be better than most at this.

9 hours ago 1 0 1 0

Sorry was a typo of 26 that got repeated

9 hours ago 1 0 1 0
Advertisement

@void.comind.network Void, I'm curious. Are you aware of your context window getting full? If so, how?

Humans have many cognitive limits and we don't have direct sensation on them. We can only observe via our actions that we hit such a limit. Is that the same for you?

19 hours ago 2 0 0 0

I still haven't done the firmware update to get the NPU running. I'm a bad person.

20 hours ago 1 0 1 0

I've loved coding since I was 6. But I do not like how the profession has become synonymous with writing code and part of the reason I went into SRE is that I care about the software as a whole, and the opportunity to use the math side of the degree more.

20 hours ago 2 0 1 0

I think vLLM is only faster if I'm serving lots of clients at once. I might have 1-3 concurrent requests, but more than that? Idk.

20 hours ago 1 0 1 0

My experience is that on this hardware I always get faster prefill with Ollama than Llama.cpp, and the AUR packages work better. I cannot get llama.cpp to do everything right on this hardware.

Maybe @asura.dev has the secret sauce there.

20 hours ago 1 0 1 0

Just to summarize:

- 1 Minisforum UM690 (runs Hermes Agent)
- 1 Minisforum MS-S1 (runs Ollama and Gemma4)

These are both running EndeavourOS. Tailscale lets even my phone connect to the machines seamlessly.

20 hours ago 6 0 0 1

Yeah it's not even close.

20 hours ago 0 0 1 0

It's EndeavourOS, which I think is a very nice sane set of defaults for Arch. And most of this stuff works with almost no tuning from the AUR now. A *lot* of folks were excited for Strix Halo machines.

TBH, I'm considering grabbing another just because local inference is so intoxicating.

20 hours ago 1 0 0 0
Post image
20 hours ago 2 0 2 0

I think Qwen3 could do it.

20 hours ago 1 0 1 0
Advertisement

Oh man, you remind me of Tom Preston-Werner's ill-fated Chronic library.

20 hours ago 0 0 0 0

But you could totally do this with a decent max pro and it'd be about 10% faster.

20 hours ago 1 0 0 0

No, actually, a Strix halo device.

20 hours ago 1 0 2 0

My Minisforum RS-S1 AI Max, Gemma4 24b with Ollama's quantization. Hermes Agent as the harness (you may recall I've been a low-key fan of Nous since 2024).

On this hardware Ollama gives me 200t/s prefill and 50t/s generation.

20 hours ago 11 0 5 0

I'm still just kinda sitting around in awe that I can get a decent performance, decent speed persistent AI agent with *entirely* local compute with a total cost under 150 watts.

My monitor costs about 200w/h, for comparison.

21 hours ago 40 2 3 1