Why aren't LLMs trained on action / cause+effect data vs. just analytical stuff?
3 points| purplerabbit | 1 year ago
Models are incredible at generating analytical / blog-ish / stack overflowish content, but suck at doing things that are complex enough that they require iteration.
For instance: If we want models that can handle complex projects, why don't we record actions taken in the execution of complex projects, and train models on that? Or if we want models that can use a browser competently, why don't we train models on screenshots + action descriptions? (Or is this what was done with o1, which is why it seems to have unprecedented capabilities?)
Is the problem just getting high-quality data? I know we've got internet dumps full of blog-ish content, but no big, easy-to-gather dumps of high-quality information about actions or chains of actions and their effects over time
(I'm sure there are tons of framing problems in this question -- sorry)
dtagames|1 year ago
I wrote an article[0] about it recently that you might enjoy.
[0] Something From Nothing | A Painless Approach to Understanding AI
https://medium.com/gitconnected/something-from-nothing-d755f...
purplerabbit|1 year ago
wmf|1 year ago