One fundamental challenge to me is that if each training run because more and more expensive, the time it takes it to learn what works/doesn't work widens. Half a billion dollars for training a model is already nuts, but if it takes 100 iterations to perfect it, you've cumulatively spent 50 billion dollars... Smaller models may actually be where rapid innovation continues simply because of tighter feedback loops. O3 may be an example of this.
ciconia|1 year ago
[1] https://hypertextbook.com/facts/2001/JacquelineLing.shtml
anon373839|1 year ago
concerndc1tizen|1 year ago
dominicrose|1 year ago
soulofmischief|1 year ago
soheil|1 year ago
dkobia|1 year ago
missedthecue|1 year ago
wruza|1 year ago
It all feels like doubling down on astrology because good telescopes aren’t there yet. I’m pretty sure that when 5 comes out, it will show some amazing benchmarks but shit itself in the third paragraph as usual in a real task. Cause that was constant throughtout gpt evolution, in my experience.
even if it kills us
Full-on sci-fi, in reality it will get stuck around a shell error message and either run out of money to exist or corrupt the system into no connectivity.
h0l0cube|1 year ago
namaria|1 year ago
soheil|1 year ago
Any technology may kill us, but we'll keep innovating as we ought to. What's your next point?
goatlover|1 year ago
idiotsecant|1 year ago
madeofpalk|1 year ago
jayseattle|1 year ago
[deleted]
khana|1 year ago
[deleted]
ulfw|1 year ago
bloodyplonker22|1 year ago
gerdesj|1 year ago
However, we are not comparing cars to horses but computers to a human.
I do want "AI" to work. I am not a luddite. The current efforts that I've tried are not very good. On the surface they offer a lot but very quickly the lustre comes off very quickly.
(1) How often do you find yourself arguing with someone about a "fact"? Your fact may be fiction for someone else.
(2) LLMs cannot reason
A next token guesser does not think. I wish you all the best. Rome was not burned down within a day!
I can sit down with you and discuss ideas about what constitutes truth and cobblers (rubbish/false). I have indicated via parenthesis (brackets in en_GB) another way to describe something and you will probably get that but I doubt that your programme will.
icpmacdo|1 year ago
https://arxiv.org/html/2410.11840v1#:~:text=Scaling%20laws%2....
merizian|1 year ago
[0] https://arxiv.org/abs/2203.03466
fny|1 year ago
falcor84|1 year ago
unknown|1 year ago
[deleted]
cma|1 year ago
From the raw scaling laws we already knew that a new base model may peter out in this run or the next with some amount of uncertainty--"the intersection point is sensitive to the precise power-law parameters":
https://gwern.net/doc/ai/nn/transformer/gpt/2020-kaplan-figu...
Later graph gpt-3 got to here:
https://gwern.net/doc/ai/nn/transformer/gpt/2020-brown-figur...
https://gwern.net/scaling-hypothesis
dyauspitr|1 year ago
soheil|1 year ago
ramesh31|1 year ago
unshavedyak|1 year ago
This also isn't true. It'll clearly have a price to run. Even if it's very intelligent, if the price to run it is too high it'll just be a 24/7 intelligent person that few can afford to talk to. No?
threeseed|1 year ago
b) There is no evidence that LLMs are the roadmap to AGI.
c) Continued investment hinges on their being a large enough cohort of startups that can leverage LLMs to generate outsized returns. There is no evidence yet this is the case.