(no title)
fergal_reid | 1 year ago
People are going to keep saying this about autoregressive models, how small errors accumulate and can't be corrected, while we literally watch reasoning models say things like "oh that's not right, let me try a different approach".
To me, this is like people saying "well NAND gates clearly can't sort things so I don't see how a computer could".
Large transformers can clearly learn very complex behavior, and the limits of that are not obvious from their low level building blocks or training paradigms.
dartos|1 year ago
Not saying I disagree with your premise that errors can’t be corrected by using more and more tokens, but this argument is weird to me.
The model isn’t intentionally generating text. The kinds of “oh let me try a different approach” lines I see are often followed by the same approach just taken. I wouldn’t say most of the time, but often enough that I notice.
Just because a model generates text doesn’t mean that the text actually represents anything at all, let alone a reflection of an internal process.
TeMPOraL|1 year ago
What does it represent then? What are all these billion weights for? It's not a bag full of NULLs that just pulls next words from a look-up table. Obviously there is some kind of internal process.
Also I don't get why people ignore the temporal aspect. Humans too generate thoughts in sequence, and can't arbitrarily mutate what came before. Time and memory is what forces sequential order - we too just keep piling on more thoughts to correct previous thoughts while they are still in working memory (context).
naasking|1 year ago
What's the mechanistic model of "intention" that you're using to claim that there is no intention in the model's operation?
> Just because a model generates text doesn’t mean that the text actually represents anything at all, let alone a reflection of an internal process.
Generating text is the trace of an internal process in an LLM.
PartiallyTyped|1 year ago
Of course, this usually requires the human to have some sense of humility and admit their mistakes.
I wonder, what if we trained more models with data that self-heals or recovers mid sentence?
yorwba|1 year ago
If the self-check is more reliable than the solution-generating process, that's still an improvement, but as long as the model makes small errors when correcting itself, those errors will still accumulate. On the other hand, if you can have a reliable external system do the checking, you can actually guarantee correctness.
solveit|1 year ago
energy123|1 year ago
mrfox321|1 year ago
Wonderfall|1 year ago
I-JEPA and V-JEPA have recently shown promising results as well.
Tostino|1 year ago