they are useful.
Posts by Yoav Goldberg
re the question about trust: I don't know if its the right question for me. getting access to sources is relevant only in a small fraction of my LLM usages (irrelevant for example when it paraphrases, styles, translates, codes, ...), and even there for me its not about trust but as anchors.
i would note that the entities who serve LLMs do have access to the training data, so they can implement your form of answer grounding, and users still do not need to know the training data composition for the LLM system to be useful.
(re attribution of claims to supported text, this does not require access to training data, only to a pile of data that contain in it texts supporting the claim, or to a search interface over such data)
they did not say "unreliable", they said incorrect sentence structure and odd word choices.
i see why transparency of data is important for research. this was not Emily's claim though. I am asking why is it important (or even needed) for users?
i think the fruitful aspects of your position on LLMs (which do exist and some of them you mention in a recent thread) would be much better served if it weren't for the non-stochastic repeated parroting of the "not useful for almost anything" mantra.
bsky.app/profile/yoav...
which usages do you find beneficial?
and why do you think they *require* transparency into the training data?
interestingness is subjective
yes, but some changes are to interesting directions and others less so
they blocked me before i got to blocking them
aaaand i am blocked
lol how i missed bsky
it is not easy and frankly the research became in many ways less fun due to LLMs, but hey they work and they are here and they are amazing and fascinating
i mean not really? yes LLMs made 95% of my technical knowledge as NLP researcher obsolete, and pretty much solved the things we were all working on. most of us said "wow this is kinda cool" and adapted.
yes. (i mean, comparatively. they perform just as well as they did back then, maybe slightly better)
(linguists mostly ignore it because it is indeed mostly irrelevant for them. yes, one could use it to learn about language, and i think it does potentially highlight some fascinating phenomena, but this is on the fringe of linguistics research and its ok.)
DL was dismissed/ignored by academia because until the 2010s or so other methods were working much better and easier to analyze (this is the ML academia, not the Linguistics/NLP academia).
DL was then adopted by NLP academia (albeit a bit late). And some weirdos like emily kept deluding themselves.
i heard it is possible to scrutinize and/or criticize a technology without being delusional
idk, sounds plausible to me that a person who actually uses the technology will understand it and its limitations much better than a person that only reads about it. dont you think?
are these *the only* things LLMs are useful for?
play this
www.puzzlescript.net/play.html?p=...
in the trailer, the mentions use the full names for "Mirror Isle" and "Skipping Stones to Lonely Homes", and "Heroes" for Heroes of Sokoban, so I'd be surprised if the intention is to hide anything.
for the record i don't think language is "solved". the parts i cared about solving, though, are to a large extent "solved", to the extent that the remaining "non-solved" parts are imo not linguistic
why?
whats the difference in your view?
i discuss this in the gist text. this is the more correct way to frame it imo (env provides observations, which agent interprets as rewards based on its goals), and it also opens up possible variations in how to think about learning from the env.
I complain a lot about RL lately, and here we go again.
The CS view of RL is wrong in how it thinks about rewards, already at the setup level. Briefly, the reward computation should be part of the agent, not part of the environment.
More at length here:
gist.github.com/yoavg/3eb3e7...
yes it sucks to be the ICLR organizers today, totally agree