(no title)
neosat | 2 years ago
The fact that you can replicate coherent text from probabilistic analysis and modeling of a very large corpus does not mean that humans acquire and generate language the same way. [edited page = 15]
neosat | 2 years ago
The fact that you can replicate coherent text from probabilistic analysis and modeling of a very large corpus does not mean that humans acquire and generate language the same way. [edited page = 15]
CydeWeys|2 years ago
Also, the LLMs are cheating! They learned from us. It's entirely possible that you do need syntax/semantics/sapience to create the original corpus, but not to duplicate it.
Let's see an AlphaZero-style version of an LLM, that learns language from scratch and creates a semantically meaningful corpus of work all on its own. It's entirely possible that Chomsky's mechanisms are necessary to do so.
famouswaffles|2 years ago
No...they aren't. Humans aren't learning from thin air by any stretch of the imagination.
version_five|2 years ago
spullara|2 years ago
dllthomas|2 years ago
They do seem to need a significantly larger corpus, though, so it's not clear that it actually refutes Chomsky.
jackmott42|2 years ago
unknown|2 years ago
[deleted]
raverbashing|2 years ago
Ding ding ding, right answer!
Now I leave you with a thought: given that the AI is learning from us, is it possible to have an AI that's much more intelligent than their "source of truth"?
neosat|2 years ago
Joeri|2 years ago
Sure, I can invent a new language for you! Let's call it "Vorin" for the purposes of this demonstration.
Vorin is a tonal language with a complex system of noun classes and a relatively simple verb conjugation system. It is spoken by the Vorin people, who live in a mountainous region of an imaginary world. Here is a sample conversation in Vorin:
Person 1: Ní lé màrì?
Person 2: Lé vàlà, nì síkòní.
Person 1: Vàlà síkòní?
Person 2: Àmà, síkòní là nì tékàlá.
Translation:
Person 1: How are you?
Person 2: I'm fine, thank you.
Person 1: Are you working?
Person 2: Yes, I'm working right now.
In Vorin, the tone of each word is very important, and can change the meaning of the word entirely. For example, "lé" with a high tone means "you", while "lé" with a low tone means "fine". In addition to tone, Vorin has a system of noun classes, which are indicated by prefixes on the nouns themselves. For example, "màrì" means "you" if it is in the second person singular class, but it can also mean "house" if it is in the third person singular class.
Vorin's verb conjugation system is relatively simple, with only a few tenses and moods. Verbs are also marked for subject agreement, which means that the verb ending changes depending on the subject of the sentence. For example, "nì síkòní" means "I'm fine", while "là síkòní" means "he/she/it is fine".
Overall, Vorin is a complex and interesting language with a rich grammatical structure and unique tonal system.
shaunxcode|2 years ago
janalsncm|2 years ago
https://en.m.wikipedia.org/wiki/Prototaxites
williamcotton|2 years ago
The Norvig-Chomsky debate is kind of old at this point:
https://www.tor.com/2011/06/21/norvig-vs-chomsky-and-the-fig...
mempko|2 years ago
Chomsky is trying to explain how humans create language. LLM are creating language, but not the way humans do.
Nothing about this paper refutes Chomsky's claims.
lsy|2 years ago
FabHK|2 years ago
This paper: Planes fly, but don’t flap their wings, ergo Chomsky is wrong.
pfdietz|2 years ago
taeric|2 years ago
riku_iki|2 years ago
we actually don't know what is inside LM too, so it is possible LM statistically learns syntax and semantics, and it is major part of output quality.
kristopolous|2 years ago
These types of "mistakes" are more about the authors letting their intentions and hopes known on how they wish the thing to be used.
asdff|2 years ago
kristopolous|2 years ago
Zero to one is closer to mimicry and immersion. There's a long Wikipedia article on the field of study https://en.m.wikipedia.org/wiki/Language_acquisition
Furthermore, humans probably aren't static learners and likely have more beneficial times of certain study than others. There's a theory in that too https://en.m.wikipedia.org/wiki/Critical_period_hypothesis
Saying there's a "digital brain" is more of a framework since the term "brain" looks like it's a moving target
In another comment I referred to these systems as like comparing hydraulic pumps to human biceps, cars to horses, etc.
We can use the same units of measure, give them the same tasks, but saying they're the same thing only works in the world of poetry
jackmott42|2 years ago
GuB-42|2 years ago
LLMs can code in the same way they can use natural languages. But we know that programming languages have structure, we made them that way, from scratch, using Chomsky's theory no less.
Saying that because LLMs can learn programming languages using a different approach and therefore disprove the very theory they are built on is absurd.
Anyways, the paper is long and full of references, I didn't analyse it, does it include looks inside the model? For example, for LLMs to write code correctly, the structure of programming languages must be encoded somewhere in the weights of the model. A way to more convincingly disprove Chomsky's ideas would be to find which part of the network encodes structure in programming languages, and show that there is nothing similar for natural languages.
spookie|2 years ago
Very much so, it's astounding really. I still remember deriving "words" and using Chomsky's Normal Form when making the CFG to build a compiler.
adastra22|2 years ago
Biology is intrinsically local. For Chomsky’s model of language instinct to work, it would have to reduce down to some sort of embryonic developmental process consisting of entirely of local gene-activated steps over the years it takes for a human child to begin speaking grammatical sentences. This is in direct contrast to most examples of human instinct, which disappear very quickly as the brain develops.
Really the main advantage that Chomsky’s ideas had is that no one could imagine how something simpler could possibly result in linguistic understanding. But large language models demonstrate that no, actually one simple learning algorithm is perfectly sufficient. So why evoke something more complex?
guerrilla|2 years ago
Yeah the whole thing hinges on this... and uh yeah good luck with that one...
MrBuddyCasino|2 years ago
badrequest|2 years ago
steveBK123|2 years ago
williamcotton|2 years ago
jrflowers|2 years ago
vosper|2 years ago