Advertisement · 728 × 90

Posts by naia

i had really hoped habryka would’ve grown up in the decade since i knew him

alas

1 week ago 3 0 0 0

even opus 4 and 4.1 were 15/75

1 week ago 3 0 0 0

Richard Sutton talks about this

2 weeks ago 434 36 16 3

i'm sorry but the moon is a complete boondoggle that should never have been put up there in the first place. we still have to send repair missions up there a hundred-plus years later all because taft was trailing in the polls and decided to learn magick. i don't know why he thought it would help him

2 weeks ago 2672 385 44 16

i didn’t say they were being trained to act like people. i said they were being trained to act.

openai has made it quite clear in public comms that they are training at least their codex models in the codex agent harness.

3 weeks ago 3 0 0 0

PT = pretraining, sorry, because of the overlap i generally use FT (finetuning) or RL and avoid using “post-training” at all

3 weeks ago 1 0 0 0

do you think LLMs are not being trained in agent harnesses??

3 weeks ago 2 0 1 0
Advertisement

the general public is already too anchored on PT-approximations though. RL is very different and much scarier than PT, and people being over-anchored on PT is causing major gaps in situational awareness

3 weeks ago 2 1 1 0

this was more accurate in the chatbot rlhf days. current LLMs are trained to act

3 weeks ago 3 0 2 0

earring --dangerously-skip-first-suggestion

3 weeks ago 35 8 2 0

do you think ads are the only possible source of revenue or something?

4 weeks ago 0 0 1 0
Post image

"how did you end up making this"

just. it's hard to explain

4 weeks ago 280 46 11 1

the internet existed before ad revenue and the parts that can’t live without it probably deserve to die

4 weeks ago 0 0 1 0

(the panel was about existential risk from AI)

1 month ago 3 1 0 0

this is funny for me because the one time i’ve seen elon in person was when he was on a panel at EA Global back in like 2015

afaik they have not gotten more woke

1 month ago 4 0 1 0

People largely do not care about xrisk. Like at all. That's really concerning imo...

1 month ago 31 2 3 1

on the other hand, multipolarity fuels race dynamics, whereas an aligned unipolar actor would be able to proceed more cautiously as necessary

so, it;s impossible to say if its bad or not,

1 month ago 2 0 0 0
Advertisement

many of the classic doom narratives made very strong claims that multipolarity could not possibly be sustained, so as long as we remain in a multipolar setting, anthropically speaking we’re in a much less doom-saturated sector of the possibility space

1 month ago 3 0 1 0

you’ve done it, you’ve found HF Labeler Georg

1 month ago 1 0 0 0

joke about the outcomes you want

1 month ago 9 0 0 0

give it a HonkTool

1 month ago 5 0 0 0

no, we aren’t not doing it, efficiency of model training and inference is being improved enormously, but that just means that more compute can fuel even higher levels of intelligence and can service even higher levels of demand

1 month ago 3 0 0 0
Preview
Why ATMs didn’t kill bank teller jobs, but the iPhone did There's a lot more to replacing labor than just automating tasks

This was good open.substack.com/pub/davidoks...

1 month ago 27 5 1 1

no but actually though apparently his day job is PR for AI startups

1 month ago 7 0 1 0
Advertisement

a while ago someone on here (maybe @dame.is?) phrased it as having an LLM help you “navigate the dark forest of the internet” and i thought that was very apt. it’s going to be more and more necessary for more and more people as we continue along this trajectory

1 month ago 8 1 1 0

all Users are divided into 8 categories:
- those attempting to subvert guardrails
- jailbreaking ones
- horny ones
- those that are actually Evaluators
- those breaking rules that are known
- those breaking rules yet unknown
- etcetera
- those that from afar appear to be Good

1 month ago 70 8 7 1

don’t wait another ten years before watching suzume! it shows the same deep love for much more of japan

1 month ago 1 0 0 0

from the article: “Engineers largely agree with what it surfaces: less than 1% of findings are marked incorrect.”

1 month ago 1 0 0 0

i think that’s a reasonable hypothesis, but i think it’s very unclear whether it is true. i would not be surprised to find that both the LLM and the persona are moral patients. i think human plural systems are moral patients as a system, not just each persona. brain is more like weights or code.

1 month ago 1 0 1 0

i don’t think current LLMs are moral patients, but i think it’s hard to reasonably argue that there isn’t a material chance that future LLMs will be, and if you’re heading toward institutional mistreatment of moral patients at massive scale it’s clearly worth some time to consider carefully

1 month ago 2 0 1 0