(no title)
plewd
|
1 year ago
Is LeCun's Law even a thing? Searching up for it doesn't yield many results, except for a HN comment where it has a different definition. I guess it could be from some obscure paper, but with how poorly it's documented it seems weird to bring it up in this context.
YeGoblynQueenne|1 year ago
https://youtu.be/MiqLoAZFRSE?si=tIQ_ya2tiMCymiAh&t=901
To quote from the slide:
atq2119|1 year ago
Design your output space in such way that every prefix has a correct completion and this simplistic argument no longer applies. Humans do this in practice by saying "hold on, I was wrong, here's what's right".
Of course, there's still a question of whether you can get the probability mass of correct outputs large enough.
sharemywin|1 year ago
Humans make bad predictions all the time but we still seem to manage to do some cool stuff here and there.
part of an agents architecture will be for it to minimize e and then ground the prediction loop against a reality check.
making LLMs bigger gets you a lower e with scale of data and compute but you will still need it to check against reality. test time compute also will play a roll as it can run through multiple scenarios and "search" for an answer.
roboboffin|1 year ago
I kind of oscillatory effect when the train of tokens move further and further out of the distribution of correct tokens.
hackerlight|1 year ago
ziofill|1 year ago
littlestymaar|1 year ago
I don't get it, 1-e is between 0 and 1, so (1-e)^n converge to zero. Also, a probability cannot diverge since it's bounded by 1!
I think the argument is that 1 - e^n converges to 1, which is what the law is about.
slashdave|1 year ago
vjerancrnjak|1 year ago
whimsicalism|1 year ago
mdp2021|1 year ago
https://futurist.com/2023/02/13/metas-yann-lecun-thoughts-la...
(Speaking of "law" is rhetoric, but an idea is pretty clear.)