(no title)
cynusx | 1 year ago
He is comparing energy spend during inference in humans with energy spend during training in LLM's.
Humans spend their lifetimes training their brain so one would have to sum up the total training time if you are going to compare it to the training time of LLM's.
At age 30 the total energy use of the brain sums up to about 5000 Wh, which is 1440 times more efficient.
But at age 30 we didn't learn good representations for most of the stuff on the internet so one could argue that given the knowledge learned, LLMs outperform the brain on energy consumption.
That said, LLM's have it easier as they are already learning from an abstract layer (language) that already has a lot of good representations while humans have to first learn to parse this through imagery.
Half the human brain is dedicated to processing imagery, so one could argue the human brain only spend 2500 Wh on equivalent tasks which makes it 3000x more efficient.
Liked the article though, didn't know about HNSW's.
Edit: made some quick comparisons for inference
Assuming a human spends 20 minutes answering in a well-thought out fashion.
Human watt-hours: 0.00646
GPT-4 watt-hours (openAI data): 0.833
That makes our brains still 128x more energy efficient but people spend a lot more time to generate the answer.
Edit: numbers are off by 1000 as I used calories instead of kilocalories to calculate brain energy expense.
Corrected:
human brains are 1.44x more efficient during training and 0.128x (or 8x less efficient) during inference.
wanderingmind|1 year ago
eloeffler|1 year ago
LLMs are always an additional cost, never more efficient because they add to the calculation, if you look at it that way.
coldtea|1 year ago
madhatter999|1 year ago
eru|1 year ago
ChatGPT has to deal with the languages we already created, it doesn't get to co-adapt.
thfuran|1 year ago
AbstractH24|1 year ago
bamboozled|1 year ago
I don't think this is true personally, ideally as children, we spend out time having fun and learning about the world is a side effect. This borg like thinking applied to intelligence because we have LLMs is unusual to me.
I learned surfing through play and enjoyment, not through training like a robot.
We can train for something with intention, but I think that is mostly a waste of energy, albeit necessary on occasion.
Jensson|1 year ago
What do you think "play" is? Animals play to learn about themselves and the world, you see most intelligent animals play as kids with the play being a simplification of what they do as adults. Human kids similarly play fight, play build things, play cook food, play take care of babies etc, it is all to make you ready for an adult life.
Playing is fun since playing helps us learn, otherwise we wouldn't evolve to play, we would evolve to be like ants that just work all day long if that was more efficient. So the humans who played around beat those who worked their ass off, otherwise we would all be hard workers.
glenstein|1 year ago
I think the part of this that resonates as most true to me is how this reframes learning in a way that tracks truth more closely. It's not all the time, 100% of the time, it's in fits and starts, its opportunistic, and there are long intervals that are not active learning.
But the big part where I would phrase things differently is in the insistence that play in and of itself is not a form of learning. It certainly is, or certainly can be, and while you're right that it's something other than Borg-like accumulation I think there's still learning happening there.
pessimizer|1 year ago
Closi|1 year ago
We don't know how to fully operate a human brain when it's fully disconnected from eyes, a mouth, limbs, ears and a human heart.
londons_explore|1 year ago
That doesn't sound right... 30 years * 20 Watts = 1.9E10 Joules = 5300 kWh.
cynusx|1 year ago
My number is based on calorie usage
CuriouslyC|1 year ago
Humans who spend a long time doing inference have not fully learned the thing being inferred - unlike LLMs, when we are undertrained, rather than a huge spike in error rate, we go slower.
When humans are well trained, human inference absolutely destroys LLMs.
cheema33|1 year ago
This isn't an apt comparison. You are comparing a human trained in a specific field to an LLM trained on everything. When an LLM is trained with a narrow focus as well, human brain cannot compete. See Garry Kasparov vs Deep Blue. And Deep Blue is very old tech.
cynusx|1 year ago
greenthrow|1 year ago
glenstein|1 year ago
I suppose they intended that as a back-of-the-envelope starting point rather than a strict claim however. But even so, gotta be accountable to your starting assumptions, and I think a lot changes when this one is reconsidered.
bufferoverflow|1 year ago
bognition|1 year ago
We probably need to exclude the cerebellum as well (which is 50% of the neurons in the brain) as it’s used for error correction in movement.
Realistically you probably just need a few parts of the lambic system. Hippocampus, amygdala, and a few of the deep brain dopamine centers.
philipov|1 year ago
pama|1 year ago
mirekrusin|1 year ago
vasco|1 year ago
marginalia_nu|1 year ago
Rinzler89|1 year ago
phantompeace|1 year ago
freehorse|1 year ago
Yes we have learnt far more complex stuff, ffs.
jryan49|1 year ago
Closi|1 year ago
i.e. not many humans invent calculus or relativity from scratch.
I think OP's point stands - these comparisons end up being overly hand-wavey and very dependent on your assumptions and view.
dist-epoch|1 year ago
So yeah, you do use 2000 calories a day, but unless you live in an isolated jungle tribe, vast amounts of energy are consumed on delivering you food, climate control, electricity, water, education, protection, entertainment and so on.
b112|1 year ago
greenthrow|1 year ago
unyttigfjelltol|1 year ago