It still can't learn. It would need to create content, experiment with it, make observations, then re-train its model on that observation, and repeat that indefinitely at full speed. That won't work on a timescale useful to a human. Reinforcement learning, on the other hand, can do that, on a human timescale. But you can't make money quickly from it. So we're hyper-tweaking LLMs to make them more useful faster, in the hopes that that will make us more money. Which it does. But it doesn't make you an AGI.
charcircuit|19 hours ago
>Reinforcement learning, on the other hand, can do that, on a human timescale. But you can't make money quickly from it.
Tools like Claude Code and Codex have used RL to train the model how to use the harness and make a ton of money.
kelnos|17 hours ago
That kind of capability is not going to lead to AGI, not even close.
Dansvidania|17 hours ago
otabdeveloper4|17 hours ago
Their contexts, not their memories. An LLM context is like 100k tokens. That's a fruit fly, not AGI.