To make this easier to discuss and think about, here's a short name <Y> for that thing, and a longer definition <Z> that probably describes that thing.
.
Oh look, if I take the literal definitions of some words in <Y>, there's a thing that fits!
> If you appeared in a puff of smoke before the authors of that paper, just after publication — a few months before half of them cleaved from OpenAI to form Anthropic — and carried with you a laptop linked through time to the big models of 2026, what would their appraisal be ? There’s no doubt in my mind they would say: Wow, we really did it ! This is obviously AGI!
I really don't think this would be the reaction. I'd say they would (or should) look at the systems we have now and see a very clear path between where they were then and where we are now, with all the positives _and negatives_. We still get hallucinations. We still get misalignment, if anything as capabilities have improved so has the potential for damage when things go wrong. It's pretty clear to me that late 2025 models are just better versions of what we had in 2021.
That's not to say they're not more useful, more valuable, they absolutely are! But that's all about product integrations, speed, and turning up the dial on inference compute. They're still fundamentally the same things.
The next big step forward, the thing that LLMs are obviously missing, is memory. The fact we're messing around with context windows, attention across the context space, chat lookup and fact saving features, etc, are all patches over the fact that LLMs can't remember anything in the way that humans (or pretty much any animal) can. It's clear that we need a paradigm shift on memory to unlock the next level of performance.
We have LLM memory, it's a training data from which the model was initially programmed. To allow adding or changing LLM memory, we would need to retrain model completely or partially. And that is not realistic any time soon. All other attempts at LLM memory would be just an obscure hack of splitting context window into parts and feeding input from different files. Literally nothing would change if you input half of the query from one file, half from another called "memory.txt" or if you just input whole query from a single file twice as big.
> It's clear that we need a paradigm shift on memory to unlock the next level of performance.
I think this is on point to the next phase of LLMs or a different neural network architecture that improves on top of them, alongside continual learning.
Adding memory capabilities would mostly benefit local "reasoning" models than online ones as you would be saving tokens to do more tasks, than generating more tokens to use more "skills" or tools. (Unless you pay more for memory capabilities to Anthropic or OpenAI).
It's kind of why you see LLMs being unable to play certain games or doing hundreds of visual tasks very quickly without adding lots of harnesses and tools or giving it a pre-defined map to help it understand the visual setting.
As I said before [0], the easiest way to understand the memory limitations with LLMs is Claude Playing Pokemon with it struggling with basic tasks that a 5 year old can learn continuously.
I'll say AGI is "here" when they are agents in our Slack/chat, emails, and planning tools doing that strategic work that management does. As a dev, "AI" is currently just a handy English language interface to web search.
Maybe AGI is here for the author and mediocre web developers, otherwise the big AI labs would have replaced their AI researchers already and commercial airliners would have already replaced their pilots with GPTs.
> This is why I propose unilateral declaration as a strategic countermove [... tearing] away the veil of any-minute-now millenarianism to reveal deployed technology
I think that in an ideal world, this would thoroughly embarrass the over-promisers by forcing them to put-up-or-shut-up, and it's fun to imagine... however I worry that it won't work out that way. Instead of deflating the nonsense in its tracks, it'll just give it more momentum and worsen the eventual mess.
> What do I mean by AGI ?
Can we fight it with a better term? Something like... Oh, I dunno, maybe "Artificial Narrative Intelligence", in the same sense that we could say A* is a kind of pathfinding intelligence.
I say "narrative" because we've got these machines that grow "fitting" documents, and are often used with stories to "decide" what happens next. For example, the story setting is a Support Page, the Customer Character says X, and the Very Helpful Robot Character then does Y and says Z in response, etc.
However just because these stories fit surprisingly well doesn't mean it's doing the kind of "thinking" we really dreamed of.
> You sometimes read about employees of AI companies absorbed by their own products. Nobody on Earth has spent more hours talking to YakGPT than Katie Echo!Nobody can pump more code out of ShannonSoft than Johnny Narcissus! Recalling my Twitter experience, I think boasts (and posts) of this kind should inspire caution.
To me a lot of that feels like just the thing-of-the-day LinkedIn Lunacy, albeit running at an unusual intensity.
I feel like it's kinda maybe here. Stochastic parrot or not, I can ask for "tea, earl grey, hot" and get an orange juice. It's way better than this time last year.
It's not perfect, but it doesn't need to be, to be useful.
tbrownaw|1 month ago
To make this easier to discuss and think about, here's a short name <Y> for that thing, and a longer definition <Z> that probably describes that thing.
.
Oh look, if I take the literal definitions of some words in <Y>, there's a thing that fits!
Therefore <X> is solved!
palmotea|1 month ago
albatross79|1 month ago
CamperBob2|1 month ago
danpalmer|1 month ago
I really don't think this would be the reaction. I'd say they would (or should) look at the systems we have now and see a very clear path between where they were then and where we are now, with all the positives _and negatives_. We still get hallucinations. We still get misalignment, if anything as capabilities have improved so has the potential for damage when things go wrong. It's pretty clear to me that late 2025 models are just better versions of what we had in 2021.
That's not to say they're not more useful, more valuable, they absolutely are! But that's all about product integrations, speed, and turning up the dial on inference compute. They're still fundamentally the same things.
The next big step forward, the thing that LLMs are obviously missing, is memory. The fact we're messing around with context windows, attention across the context space, chat lookup and fact saving features, etc, are all patches over the fact that LLMs can't remember anything in the way that humans (or pretty much any animal) can. It's clear that we need a paradigm shift on memory to unlock the next level of performance.
Yizahi|1 month ago
rvz|1 month ago
I think this is on point to the next phase of LLMs or a different neural network architecture that improves on top of them, alongside continual learning.
Adding memory capabilities would mostly benefit local "reasoning" models than online ones as you would be saving tokens to do more tasks, than generating more tokens to use more "skills" or tools. (Unless you pay more for memory capabilities to Anthropic or OpenAI).
It's kind of why you see LLMs being unable to play certain games or doing hundreds of visual tasks very quickly without adding lots of harnesses and tools or giving it a pre-defined map to help it understand the visual setting.
As I said before [0], the easiest way to understand the memory limitations with LLMs is Claude Playing Pokemon with it struggling with basic tasks that a 5 year old can learn continuously.
[0] https://news.ycombinator.com/item?id=43291895
fragmede|1 month ago
akagusu|1 month ago
gedy|1 month ago
denkmoon|1 month ago
andsoitis|1 month ago
Can AGI not speak for itself? Does it need humans to speak and act on its behalf? Who are the high-priests and what are the sects?
smitty1e|1 month ago
rvz|1 month ago
Maybe AGI is here for the author and mediocre web developers, otherwise the big AI labs would have replaced their AI researchers already and commercial airliners would have already replaced their pilots with GPTs.
This is exactly why “AGI” is meaningless.
ares623|1 month ago
unknown|1 month ago
[deleted]
Terr_|1 month ago
> This is why I propose unilateral declaration as a strategic countermove [... tearing] away the veil of any-minute-now millenarianism to reveal deployed technology
I think that in an ideal world, this would thoroughly embarrass the over-promisers by forcing them to put-up-or-shut-up, and it's fun to imagine... however I worry that it won't work out that way. Instead of deflating the nonsense in its tracks, it'll just give it more momentum and worsen the eventual mess.
> What do I mean by AGI ?
Can we fight it with a better term? Something like... Oh, I dunno, maybe "Artificial Narrative Intelligence", in the same sense that we could say A* is a kind of pathfinding intelligence.
I say "narrative" because we've got these machines that grow "fitting" documents, and are often used with stories to "decide" what happens next. For example, the story setting is a Support Page, the Customer Character says X, and the Very Helpful Robot Character then does Y and says Z in response, etc.
However just because these stories fit surprisingly well doesn't mean it's doing the kind of "thinking" we really dreamed of.
> You sometimes read about employees of AI companies absorbed by their own products. Nobody on Earth has spent more hours talking to YakGPT than Katie Echo! Nobody can pump more code out of ShannonSoft than Johnny Narcissus! Recalling my Twitter experience, I think boasts (and posts) of this kind should inspire caution.
To me a lot of that feels like just the thing-of-the-day LinkedIn Lunacy, albeit running at an unusual intensity.
shmerl|1 month ago
tra3|1 month ago
It's not perfect, but it doesn't need to be, to be useful.
petermcneeley|1 month ago
lostmsu|1 month ago