Isn’t this like a brute force approach?
Given it costs $ 3000 per task, thats like 600 GPU hours (h100 at Azure)
In that amount of time the model can generate millions of chains of thoughts and then spend hours reviewing them or even testing them out one by one. Kind of like trying until something sticks and that happens to solve 80% of ARC. I feel like reasoning works differently in my brain. ;)
tikkun|1 year ago
TrapLord_Rhodo|1 year ago
the LLM only gets two guesses at the "end solutions". The whole chain of thought is breaking out the context, and levels of abstraction. How many "Guesses" is it self generating and internally validating, well that's all just based on compute power and time.
My counter point to OP here would be is that is exactly how our brain works. In every given scenario, we are also evaluating all possible solutions. Our entire stack is constantly listening and eithier staying silent, or contributing to an action potential (eithier excitatory, or inhibitory). but our brain is always "Evaluating all potential possibilities" at any given moment. We have a society of mind always contributing their opinion, but the ones who don't have as much support essentially get "Shouted down".
nmca|1 year ago
unknown|1 year ago
[deleted]
macrolime|1 year ago
nextworddev|1 year ago
Its not agi obviously in the sense that you still need to some problem framing and initialization to kickstart the reasoning path simulations
torginus|1 year ago
strangescript|1 year ago
"Well, yeah, but its kind of expensive" -- this guy
tymonPartyLate|1 year ago
jeremyjh|1 year ago
freehorse|1 year ago