(no title)
oldge
|
7 months ago
Today’s llms are fancy autocomplete but lack test time self learning or persistent drive.
By contrast, an AGI would require:
– A goal-generation mechanism (G) that can propose objectives without external prompts
– A utility function (U) and policy π(a│s) enabling action selection and hierarchy formation over extended horizons
– Stateful memory (M) + feedback integration to evaluate outcomes, revise plans, and execute real-world interventions autonomously
Without G, U, π, and M operating llms remain reactive statistical predictors, not human level intelligence.
KoolKat23|7 months ago
Looking at the human side, it takes a while to actually learn something. If you've recently read something it remains in your "context window". You need to dream about it, to think about, to revisit and repeat until you actually learn it and "update your internal model". We need a mechanism for continuous weight updating.
Goal-generation is pretty much covered by your body constantly drip-feeding your brain various hormones "ongoing input prompts".
onemoresoop|7 months ago
How are we not far off? How can LLMs generate goals and based on what?
NetRunnerSu|7 months ago
https://github.com/dmf-archive/PILF
asah|7 months ago
unknown|7 months ago
[deleted]
NetRunnerSu|7 months ago
https://dmf-archive.github.io/docs/posts/beyond-snn-plausibl...