(no title)
mo_42 | 1 year ago
I'm wondering when people will apply this to other areas like the real world. Would it learn the game engine of the universe (ie physics)?
mo_42 | 1 year ago
I'm wondering when people will apply this to other areas like the real world. Would it learn the game engine of the universe (ie physics)?
radarsat1|1 year ago
I think for real world application one challenge is going to be the "action" signal which is a necessary component of the conditioning signal that makes the simulation reactive. In video games you can just record the buttons, but for real world scenarios you need difficult and intrusive sensor setups for recording force signals.
(Again for robotics though maybe it's enough to record the motor commands, just that you can't easily record the "motor commands" for humans, for example)
cubefox|1 year ago
https://slatestarcodex.com/2017/09/05/book-review-surfing-un...
It's called predictive coding. By trying to predict sensory stimuli, the brain creates a simplified model of the world, including common sense physics. Yann LeCun says that this is a major key to AGI. Another one is effective planning.
But while current predictive models (autoregressive LLMs) work well on text, they don't work well on video data, because of the large outcome space. In an LLM, text prediction boils down to a probability distribution over a few thousand possible next tokens, while there are several orders of magnitude more possible "next frames" in a video. Diffusion models work better on video data, but they are not inherently predictive like causal LLMs. Apparently this new Doom model made some progress on that front though.
ccozan|1 year ago