(no title)
gateorade | 2 years ago
The type of work I do is highly niche. I’ve recently been working on a specific problem for which there are probably only a hundred at most implementations running on production systems, all of them highly proprietary. I would be surprised if there were any implementations in GPTs training set. With that said, this problem is not actually that complicated. A rudimentary implementation can be done in ~100 lines of code.
I asked GPT-4 to write me an implementation. It knew a decent amount about the problem (probably from Wikipedia). If it was actually capable of something close to reasoning it should have been able to write an implementation, but when it actually started writing code it was reluctant to write more than a skeleton. When I pushed it to implement specific details it completely fell apart and started hallucinating. When I gave it specific information about what it was doing wrong it acknowledged that it made a mistake and simply gave me a new equally wrong hallucination.
The experience calmed my existential fears about my job being taken by AI.
softfalcon|2 years ago
He was convinced that if we trained the AI on enough data, GPT-x would become sentient.
My opinion was similar to yours. I felt like the hallucinating the AI does was insufficient in performing true extrapolating thought.
I said this because humans don’t truly have access to infinite knowledge, even when they do, they can’t process all of it. Adding endless information for the AI to feed on doesn’t seem like the solution to figuring out true intelligence. It’s just more of the same hallucinating.
Yet despite lacking knowledge, us humans still come up with consistently original thoughts and expressions of our intelligence daily. With limited information, our minds create new representations of understanding. This seems to be impossible for Chat GPT.
I could be completely wrong, but that discussion solidified for me that my role as a dev still has at least a couple more decades of shelf life left.
It’s nice to hear that others are reaching similar conclusions.
visarga|2 years ago
>> Write a response that includes the number of words in your response.
> This response contains exactly sixteen words, including the number of words in the sentence itself.
It contains 15 words.
The model would have to plan everything before outputting the first token if it were to solve the task correctly. Works if you follow up with "Explicitly count the words", let it reply, then "Rewrite the answer".
andsoitis|2 years ago
It turns out it isn’t just AIs that hallucinate; AI researchers do as well.
Majromax|2 years ago
Is there enough data?
As I understand it, the latest large language models are trained on almost every piece of available text. GPT-4 is multimodal in part because there isn't an easy way to increase its dataset with more text. In the meantime, text is already quite information dense.
I'm not sure that future models will be able to train on an order of magnitude more information, even if the size of their training sets has a few more zeroes added to the end.
psychphysic|2 years ago
So he might be right but due to time and not due to improved performance.
I believe in the UK all vertibrates are considered sentient (by law not science). That includes goldfish.
And good luck even getting a goldfish to reverse a linked list. Even after 1000 implementations are provided.
wslh|2 years ago
Not saying your friend is right or wrong, but imagine if civilization gives more information, in realtime, to an AI system through sensors: will be at least sentient as the civilization? Seems like a scifi story, a competitor to G-d.
kaba0|2 years ago
Correct me if I’m wrong, but doesn’t it mean that they can’t recursively “think”, on a fundamental basis? And sure I know that you can pass “show your thinking” to GPT, but that’s not general recursion, just “hard-coded to N iterations” basically, isn’t it? And thus no matter how much hardware we throw at it, it won’t be able to surpass this fundamental limit (and without proof, I firmly believe that for a GAI we do need the ability to basically follow through a train of thought)
aiphex|2 years ago
dmichulke|2 years ago
I think there is some sampling bias in your observation ;-)
oliveiracwb|2 years ago
antonvs|2 years ago
It sounds like he doesn't even understand the basics of what GPT is, or what sentience is. GPT is an impressive manipulator/predictor of language, but we have evidence from all sorts of directions that there's more to sentience or consciousness than that.
suction|2 years ago
[deleted]
braindead_in|2 years ago
vasco|2 years ago
The issue is that among all the 100k+ software engineers, many don't really do anything novel. How many startups are employing dozens of engineers to create online accessible CRUDs to replace a spreadsheet?
In the company I work for I'd say we have about 15 developers or about 3 teams doing interesting work, and everyone else builds integrations, CRUDs, moves a button there and back in "an experiment", ads a new upsell, etc. All these last parts could be done by a PM or good UX person alone, given good enough tools.
The other parts I'm not worried about either.
yohannesk|2 years ago
oblio|2 years ago
Figuring out what code to write is one of the big parts.
Fixing it when it breaks in many creative ways is the other big part.
How good is ChatGPT at fixing bugs? Security bugs or otherwise?
sterlind|2 years ago
It just spit out garbage. Because (afaict) there aren't really examples of that specific thing on the Internet. And it's just been weirdly bad at all the cartography-related programming problems I've thrown at it, in general.
And yeah, I'm much less worried about it replacing me now. It's just not.. lucid, yet.
laurels-marts|2 years ago
noduerme|2 years ago
v4dok|2 years ago
It might not know more than you about your niche. I don't. I would search and I would try to reason, but if I was forced to give a token by token output that is answering the question as truthfully as possible, I might have started saying bullshit as well.
I don't think that the fact that gpt doesn't know things or does some things wrong is sufficient to save dev work from automation.
[1]: https://github.com/noahshinn024/reflexion-human-eval
blablabla123|2 years ago
Same for me. I didn't try GPT-4 yet, and not on code from work anyway but GPT-3 seems borderline useless at this point. The hallucinations are quite significant. Also I tried to produce advice for Agile development with references and as stated in other articles the links where either 404s or even completely unrelated articles.
Still I'm taking this seriously. Just considering the leaps that happened with AlphaGo/AlphaZero or autonomous driving, that was considered unthinkable in the respective domains before.
zeroonetwothree|2 years ago
After all, just look at manufacturing. Compared to 1970 we produce 5x the real output but employ only 50% the people. The same will likely happen to fields like programming as AI improves.
olivermuty|2 years ago
nimbix|2 years ago
Event after several iterations of giving it error messages and writing explanations of what's not working, it didn't even get past the first issue. Sometimes it would agree that it needs to fix something, but would then print back code with exactly the same problem.
toss1|2 years ago
I wrote some questions in the specialist legal field of someone in my household, then started to get into more specialist questions, and then specifically asked about a paper that she wrote innovating a new technique in the field.
The general question answers were very impressive to the attny. The specialist questions started turning up errors and getting concepts backwards - bad answers.
When I got to summarizing the paper with the new technique, it could not have been more wrong. It got the entire concept backwards and wrong, barfing generic and wrong phrases, and completely ignored the long list of citations.
Worse yet, to the point of hilariously bad, when asked for the author, date, and employer of the paper, it was entirely hallucinating. Literally, the line under the title was the date, and after that was "Author: [name], [employer]". It just randomly put up dates and names (or combinations of real names) of mostly real authors and law firms in the region. Even when pointed out the errors, it would apologize, and then confidently spout a new error. Eventually it got the date correct, and that stuck, but even when prompted with "Look at where it says 'Author: [fname]" and tell me the full name and employer, it would hallucinate a last name and employer. Always with the complete confidence of a drunken bullshit artist.
Similar for my field of expertise.
So, yes, for anything real, we really need to keep it in the middle-of-the-road zone of maximum training. Otherwise, it will provide BS (of course if it is BS we want, it'll produce it on an industrial scale!).
willbudd|2 years ago
funstuff007|2 years ago
This is true to varying degrees for every statistical model ever.
gateorade|2 years ago
lostmsu|2 years ago
SketchySeaBeast|2 years ago
mannykannot|2 years ago
m3kw9|2 years ago