top | item 45455936

(no title)

srj | 5 months ago

Yes I'm talking about LLMs in particular. I'm in the stochastic parrot camp. Though I could be convinced humans are no more than stochastic parrots, in which case it does have a path for development of AGI.

If I'm right the breakthroughs will plateau even while applications of the technology continue to advance for the next several years.

discuss

order

xpe|4 months ago

Here is my take. When people use the stochastic parrots phrase, very often they use it as an explanation of what is happening. But in many cases, I don't think they appreciate: (1) good explanations must be testable models; (2) different explanations exist at different levels of abstraction; (3) having one useful level of explanation does not mean that other levels of explanation are not accurate nor useful.

Sure, optimization based on predicting the next word is indeed the base optimizer for LLMs. This doesn't prevent the resulting behavior from demonstrating behavior that corresponds with some measurable levels of intelligence, as in problem-solving in particular domains! Nor does it prevent fine-tuning from modifying the LLMs behavior considerably.

One might say e.g. "LLMs only learn to predict the next word." The word only is misleading. Yes, models learn to predict the next word, and they build a lot of internal structures to help them do that. These structures enable capabilities much greater than merely parroting text. This is a narrow claim, but it is enough to do serious damage to the causal wielder of the "stochastic parrots" phrase. (To be clear, I'm not making any claims about consciousness or human-anchored notions of intelligence.)

xpe|4 months ago

> I'm in the stochastic parrot camp.

If you wouldn't mind doing me a favor?... For a few minutes, can we avoid this phrase? I don't know what people really mean by it.* Can you in plain English translate your view into sentence of theses form:

1. "Based on my understanding of LLMs, X1 and X2 are impossible."

2. "This leads to predictions such as P1 and P2."

3. "If I observed X3 or X4, it would challenge my current beliefs of LLMs."

* I've read the original stochastic parrots paper. In my view, the paper does not match how many people talk about it. Quite the opposite. It is likely many people name drop it but haven't read it carefully. I may have some misinterpretations, sure, but at least I'm actively seeking to question them.