(no title)
charcircuit | 18 hours ago
>Reinforcement learning, on the other hand, can do that, on a human timescale. But you can't make money quickly from it.
Tools like Claude Code and Codex have used RL to train the model how to use the harness and make a ton of money.
kelnos|16 hours ago
That kind of capability is not going to lead to AGI, not even close.
regularfry|12 hours ago
1. It's still memory, of a sort, which is learning, of a sort. 2. It's a very short hop from "I have a stack of documents" to "I have some LoRA weights." You can already see that happening.
charcircuit|15 hours ago
One of the biggest boosts in LLM utility and knowledge was hooking them up to search engines. Giving them the ability to query a gigantic bank of information already has made them much more useful. The idea that it can't similarly maintain its own set of information is shortsighted in my opinion.
Dansvidania|16 hours ago
regularfry|12 hours ago
otabdeveloper4|17 hours ago
Their contexts, not their memories. An LLM context is like 100k tokens. That's a fruit fly, not AGI.
charcircuit|15 hours ago