(no title)
durovo
|
3 years ago
I believe GPT-3 has a transformer-based architecture. So it doesn't recursively ingest it's own output in each iteration. I believe attention-based transformer models have enough complexity to be able to learn what you are talking about on their own.
ravi-delia|3 years ago
thrtythreeforty|3 years ago