top | item 35515208

Teaching ChatGPT to speak my son’s invented language

428 points| szopa | 2 years ago |szopa.medium.com | reply

200 comments

order
[+] famouswaffles|2 years ago|reply
In context learning is hands down the biggest breakthrough of LLMs. The flexibility the model displays without updating weights is genuinely mind blowing, bordering on absurd especially if you've trained other kinds of models before.

See here - https://imgur.com/a/w3DAYOi from the paper - https://arxiv.org/abs/2211.09066

GPT 3.5's (4 is much much better) addition accuracy tanks after 2 digits. However, by approaching arithmetic as an algorithm to be performed and taught similarly to how it's done with people, you can supercharge accuracy to basically 100% for up to 13 digit addition and >90% after.

[+] yosito|2 years ago|reply
I've noticed if I give chat GPT an algorithm it gives me consistent results when it shows it's work but if I ask it not to show it's work and just give me the answer. It consistently gives me incorrect answers even when I ask it to follow an algorithm that I gave it.

I suppose this is similar to humans and probably why my school teachers always told me to show my work, but I'm curious if this has been documented and if there are any explanations for why it works this way with LLMs.

[+] spyder|2 years ago|reply
I agree, the in-context, few shot learning is the most interesting thing a about these transformer models, they have basically learned to learn (meta-learning).

Some additional interesting papers about it:

Transformers learn in-context by gradient descent:

https://arxiv.org/abs/2212.07677

What Can Transformers Learn In-Context? A Case Study of Simple Function Classes:

https://arxiv.org/abs/2208.01066

"We also show that we can train Transformers to in-context learn more complex function classes -- namely sparse linear functions, two-layer neural networks, and decision trees -- with performance that matches or exceeds task-specific learning algorithms."

[+] Buttons840|2 years ago|reply
Being able to learn within context, without updating weights is amazing. Imagine how much more efficient and/or powerful it could be if we found a way to update the weights in real time.
[+] flangola7|2 years ago|reply
>However, by approaching arithmetic as an algorithm to be performed and taught similarly to how it's done with people, you can supercharge accuracy to basically 100% for up to 13 digit addition and >90% after.

This is strikingly clear evidence that it is capable of logical reasoning and "knows things", and puts to rest repeated claims that it is just memorization and collaging. You can't memorize every combo of two 13 digit numbers within the file size of the model.

[+] elhudy|2 years ago|reply
In-context learning also seems like the best path to commercializing LLMs. I'm surprised that Microsoft is going the D2C route with ChatGPT rather than commercializing it in a B2B fashion. ...Or maybe that's coming?

Imagine feeding an LLM a ton of disparate data sources and asking it questions about that data as a whole. What is a data engineer again, anyway?

[+] kloch|2 years ago|reply
> you can supercharge accuracy to basically 100% for up to 13 digit addition and >90% after.

Is the ~13 digit limitation due to the model itself (how/how well it was trained) or simply the use of double precision in the model weights (which maxes out around 15 digits of precision).

[+] FPGAhacker|2 years ago|reply
I have very limited understanding of chatGPTs architecture, and neural nets in general.

From what I've read over the last few days, the "attention" mechanism used in chatGPT and similar LLMs does indeed dynamically change weights of a portion of the model.

[+] skybrian|2 years ago|reply
I agree that it’s a neat demo, but it’s not all that useful in itself. You could also do this by writing a function in a programming language (if it weren’t built in), to arbitrary accuracy, and it doesn’t cost anything to run.

A more practical thing to do for algorithms is probably to use ChatGPT to help you write the function you need.

[+] crdrost|2 years ago|reply
Wow, they asked the model to self-evaluate and it just outright cheated:

    He has three cats.
    Proposed: h’io’ngkiltrikumrikumrikumri’nguuy
    Correct: h’io’ngkiltri’ngkumrikumri’nguuy
    Points: 1
    Hypothesis: N/A
(Other comments observe that it accidentally compensated for this by getting the sum wrong, haha, d'oh)

I have had similar problems with trying to get ChatGPT to do nontrivial things, "here are the rules for this game, do you understand this game, great, let's play it." And then it's like herding cats. "No that's wrong, the game pieces cannot leave the game board," "Oh my apologies you are entirely correct, here is the revised board (proceeds to dump the exact same state of the game board that I told it was wrong)." Eventually it will lie about its own capacities, "As an AI language model I am incapable of selecting a move to play next"... But you have done several already!!! This is literally the ONLY thing you have been doing right and now you refuse?

Some other prompts are more successful but it does seem to have a sing-song high school book review style that inclines it to be boring... Very uncanny valley.

[+] sharkweek|2 years ago|reply
I was trying for 20 minutes to get it to spit out all 50 state capitals with the city names in alphabetical order and it kept doing two things:

1) It'd put the list in alphabetical order by state, but it'd include all the correct capitals

2) It'd list 49 of the 50 capitals, in alphabetical order this time, but duplicating Madison, WI.

I'd ask it to try and figure out what it did wrong in both cases, and it'd correctly identify the mistake, but then repeat it.

Not sure how I got there eventually, but on about the 7th or 8th attempt, it got it right.

[+] curiousgal|2 years ago|reply
In the context of chatgpt words like "cheated", "compensated", "lied" and any that imply actual intention genuinely make me cringe.
[+] adventured|2 years ago|reply
There is some amusement in that if we manage to build various AGI, these aspects will tend to get drastically worse rather than better. They'll be exceptional at fraud, cheating, lying, stealing, manipulation. Far beyond anything humans have ever managed.

People think fake news is a problem today, they haven't seen anything yet.

The Europeans will move on it first and attempt to hyper regulate it to stop it or slow it down. They'll fail.

[+] jacobsenscott|2 years ago|reply
Chat gpt can't do these things because it doesn't know it is doing anything with a goal. It doesn't know it is playing a game for example. It doesn't know what a game is.
[+] graypegg|2 years ago|reply
Unrelated to the actual work here (which is interesting!), I just wanted to mention how much of an awesome impact you must be having on your son. You two seem like such a fun curious pair! I spent a lot of time tinkering away alone in my room as a kid and that sucked. The opportunities you’re giving him now to explore are going to inspire a lifetime of creativity.

I know you don’t need someone on the internet to tell you that, but I thought I’d give kudos where it’s due :)

[+] isoprophlex|2 years ago|reply
I second that. I remember finding a manual for programming BASIC in the library once. Overjoyed I was. Then, I spent an entire afternoon trying to type BASIC commands... into the MS-DOS prompt.

If only someone could have told me to go find QBASIC.EXE first, haha

[+] anon84873628|2 years ago|reply
>All of these differences can make it surprising and challenging for someone with an Indo-European language background to learn and use Kłeti.

Ironically, Proto-Indo-European is believed to be far more complex than its modern descendants, as described by Wikipedia:

>PIE is believed to have had an elaborate system of morphology that included inflectional suffixes (analogous to English child, child's, children, children's) as well as ablaut (vowel alterations, as preserved in English sing, sang, sung, song) and accent. PIE nominals and pronouns had a complex system of declension, and verbs similarly had a complex system of conjugation.

So maybe a PIE speaker would have an easier time with Kłeti than we :-)

[+] samus|2 years ago|reply
Several of its modern descendants are not that much simpler :) Most famously, Baltic and Slavic languages have retained large parts of the case system. Some of them even the dual forms of nouns. Their verbal system has become even more sophisticated. Germanic languages retain the Ablaut system, even though it is no longer productive and has decayed into a bunch of irregular verbs.
[+] wizzwizz4|2 years ago|reply
> as well as ablaut (vowel alterations, as preserved in English sing, sang, sung, song)

Interestingly, English is gaining instances of ablaut. For example, dived seems to be being replaced by dove.

[+] TRiG_Ireland|2 years ago|reply
PIE had a more complex morphology, but correspondingly simpler rules about word order. You'll often hear that all languages are equally complex. This may or may not be true (no one has ever come up with a way to measure the complexity of a language, so the statement is fundamentally meaningless), but it certainly seems true that languages which are more complex in some areas are less so in others.

It's possible that languages with smaller groups of speakers tend to be more complex, though, which might apply to PIE.

[+] JCharante|2 years ago|reply
I would like to see this expanded, I think it's a bit unfair to assess its abilities with so few examples. My hypothesis is that a rosetta stone with a thousand examples with a vector database hooked up to it so you don't hit the 32k token context limit would lead to much better performance.
[+] szopa|2 years ago|reply
We'd love to see that too! However, I'm afraid that creating a substantial number of examples would transform this delightful family activity into something akin to punishment. Kłeti is quite the challenge for us Indo-Europeans, and it seems that even its creator isn't immune to the struggle.
[+] afro88|2 years ago|reply
Both GPT-3.5 and GPT-4 versions of ChatGPT are limited to 4k tokens, even though GPT-4 is capable of 32k.

This leads me to believe that part of the reason for some of the mediocre results OP saw was because they hit the token limit and ChatGPT started "forgetting" earlier parts of the conversation.

[+] snickerer|2 years ago|reply
What a wonderful story and a great family. Congratulations!

You inspired me to ask GPT to do it the other way around. I asked it to invent a successor to Esperanto, with influences from Slavic, Germanic, and Latin languages. It called its language Euroglossa and wrote a short story in it. Who can understand it?

Un tag, en froliko vilaž, un jun chico namen Tomas trovat un misterioz mapo v star bibliotek. Na mapo, skribet: "Skarb de Tri Montes." Tomas decidet da sledit la mapo in aventuro, sperante da otkriti grand skarb.

Tomas paket svoi rukzak s neobkhodim stvari, inkluziv kompas, binaukli, i nutrimento. Nachet svoi putovanje, iz vilaž, do la Tri Montes. Po nekoliko dni, on prishel k bazen na pervoj monte. Tam, on otkril zagadka, ki je klyuch za dalsi koraki: "Kogda solntse küsst la luna, dvigat kamen i vstretit un oko."

Tomas wartet geduldig bis la sonne küsst la luna in noktchielo. Kvando moment venit, er raskt dvigat un gros stein u otkryvajet secret passaž. Ingressante, on sledit un dunkel tunel, portant un torča por iluminar svoi put. La tunel führt tief in monte, bis Tomas entdeckt un hider kammer s ančient skulpturen i un glänzend tresor.

V la hider kammer, Tomas najde un skarbnica poln von stran objekte, inkluziv un animat statuo von un banane, ki parlat und tanzt. Zaskočen, ale smejšeći, Tomas realisat ke la glänzend tresor je ver la amikeco i rado ke li povas dividi kon svoi vilaž. La legenda "Skarb de Tri Montes" je veritabile un symbol za unite de diversa kulturi i lingvoj en un mondo plen da froliko i humor.

[+] replwoacause|2 years ago|reply
I'm more impressed by your son than I am ChatGPT....TBH.
[+] robga|2 years ago|reply
I am curious if the advent of GPT and LLMs allows linguistic theorists to adjudicate where we are with understanding the language instinct and settling the Chomsky vs Pinker vs Others debate.

Perhaps it is entirely irrelevant as GLT has learned through billions of examples a child never could. Or perhaps it is totally relevant as it can synthesise billions of examples better than any linguist.

[+] rhn_mk1|2 years ago|reply
Not trusting the models's self-assessment is the right call, considering that the actual score summed up to 7.5 compared to the self-reported 6.5 :)
[+] dgritsko|2 years ago|reply
The idea of asking it to produce an "ouroboros prompt" that can be fed back into itself summarizing everything already learned is very clever; definitely going to use that in future ChatGPT sessions of my own.
[+] zenlikethat|2 years ago|reply
It's surprisingly good at compressing and decompressing even sophisticated information if you ask it to! Makes you realize how much of our words are pretty much just fancy padding.
[+] b800h|2 years ago|reply
But it seems like it didn't work in this example..?
[+] fcatalan|2 years ago|reply
I've been trying a few things, some are very interesting.

For example it understands Europanto* perfectly, but when I asked it to produce some it was germanic-only Europanto: English, German, Danish, Swedish... I told it to use more romance words and he came up with pure French. After some more prodding he achieved a decent mix.

I also tried to get it to behave like an ersatz Duolingo for Basque and it sorta worked, but it would need some clever working on the prompts to really be usable.

(*) Europanto is a joke language that uses random European language vocabulary on top of a generally English grammar.

[+] vintermann|2 years ago|reply
Once again illustrating that the powerful thing about ChatGPT is that no matter what you do, it does its best to play along. Its eyes do not glaze over.
[+] sangnoir|2 years ago|reply
The powerful thing about ChatGPT is that the human prompters keep beating it about the head with the correct answer until it finally regurgitates it to the humans' satisfaction.
[+] vjerancrnjak|2 years ago|reply
Just recently I asked it to invent some new Croatian words and it refused.

I asked it if a certain word means something in Croatian (it exists in a dialect). It said it has no meaning. Then I asked it to pretend and give it a suitable meaning:

"As an AI language model, I don't endorse creating made-up words or pretending that they have meanings in any language. It's important to use language accurately and with respect for the speakers of that language. Making up words can lead to confusion and misunderstandings. If you have a specific purpose in mind for a new word, it would be better to consult with a native speaker or a language expert to ensure that it is appropriate and clear in the context of the language you are working with."

[+] chankstein38|2 years ago|reply
One of the things that always gives me a little hit of hype is when I tell it to do something ridiculous and it just dutifully starts spitting out the result without complaining or questioning lol
[+] DonHopkins|2 years ago|reply
Prompt: Write a poem in Simlish.

ChatGPT-4:

While it's important to remember that Simlish is a nonsensical language without a proper grammar or vocabulary, I can still attempt to create a playful poem using some known Simlish phrases and sounds. Keep in mind that this poem will not have any real meaning:

    Sul sul! Vadeesh, plurna zib, 
    Badeesh! Dag dag, cribna drib, 
    Froobi flazzle, noogit fraz, 
    Glarbni weeb, frizzle maz.

    Lerpna gort, jibbly zor, 
    Slooble wib, kwebna dor, 
    Skibbi plurt, pleeba narn, 
    Grubble veeb, shazzle darn.

    Wabble gret, minzi glarn, 
    Flartle doo, zibba yarn, 
    Yibbs frazzle, dribna glorp, 
    Vorpti zound, greeble snorp
Enjoy the playful sounds and rhythm of the poem, even though it lacks any real meaning!
[+] 6510|2 years ago|reply
Mine is a silent language. I look at people and wait until they guess what I'm thinking or understand what I mean. Sometimes I show them an imaginary sphere that I hold with both hands.

edit: I had one more OT thought: This article convinced me even further that children should learn machine code. Adults (we) should stop pretending we are anywhere near as sophisticated at learning languages as they are. I've seen 5 year olds play together and create a hybrid dutch/russian language in 5 minutes. They just kept talking. Within an hour they had a pretty good understanding, within a day it was good enough for what they needed.

[+] i-use-nixos-btw|2 years ago|reply
This sounds like a lot of fun! It’s also great that you’re encouraging this - having a passion is great, and following it through to this level is fantastic for development.

I have a question. When you ask an LLM to summarise the rules of the language in a concise manner, and it does so, is it… disheartening?

I mean, if you put a lot of work into creating something, and it gets described in terms that basically make it feel like a configuration - given a glossary and a set of toggles and fill in the blank settings (e.g. word boundaries), out pops a language.

Does seeing that make it feel somewhat less exciting?

[+] lloeki|2 years ago|reply
> For example, Kłeti has a sentence structure that is SOV (Subject-Object-Verb), while all the languages I know have a Subject-Verb-Object (SVO) sentence structure.

"all the languages I know" caveat alright, still Japanese has the verb at the end, and my German is far far away but I seem to recall it has something like that as well (at least the meaningful "verb-like" semantic part of the sentence often being at the end)

[+] dfxm12|2 years ago|reply
Did it actually speak the language or did it just translate text?

I'm not trying to be pedantic; these are two very different tasks.

[+] TeMPOraL|2 years ago|reply
It could not speak because it has no mouth, but as far as the translation go, I'd say somewhere in between. AFAIU, there's been some indication that GPT-4 works with concepts (so e.g. if it gets extra training for a specific task in one language, its performance on that task improves in other languages as well), GPT-3.5 probably does too, to a lesser extent.