I think it's a lot more complicated than that. I think it can be used as a tool for people who already have knowledge and skills, but I do worry how it will affect people growing up with it.
Personally I see it more like going to someone who (claims) to know what they're doing and asking them to do it for me. I might be able to watch them at work and maybe get a very general idea of what they're doing but will I actually learn something? I don't think so.
Now, we may point to the fact that previous generations railed at the degeneration of youth through things like pocket calculators or mobile phones but I think there is a massive difference between these things and so-called AI. Where those things were tools obligatorily (if you give a calculator to someone who doesn't know any formulae it will be useless to them), I think so-called AI can just jump straight to giving you the answer.
I personally believe that there are necessary steps that must be passed through to really obtain knowledge and I don't think so-called AI takes you through those steps. I think it will result in a generation of people with markedly fewer and shallower skills than the generations that came before.
AI will let some people conquer skills otherwise out of their reach, with all the pros and cons of that. It is exactly like the example someone else brought up of not needing to know assembly anymore with higher level languages: true, but those who do know it and can internalize how the machines operate have an easier time when it comes to figuring out the real hard problems and bugs they might hit.
Which means that you only need to learn machine language and assembly superficially, and you have a good chance of being a very good programmer.
However, where I am unsure how the things will unfold is that humans are constantly coming up with different programming languages, frameworks, patterns, because none of the existing ones really fit their mental model or are too much to learn about. Which β to me at least β hints at what I've long claimed: programming is more art than science. With complex interactions between a gazillion of mildly incompatible systems, even more so.
As such, for someone with strong fundamentals, AI tools never provided much of a boon to me (yet). Incidentally, neither did StackOverflow ever help me: I never found a problem that I struggled with that wasn't easily solved with reading the upstream docs or upstream code, and when neither was available or good enough, SO was mostly crickets too.
These days, I rarely do "gruntwork" programming, and only get called in on really hard problems, so the question switches to: how will we train the next generation of software engineers who are going to be called in for those hard problems?
Because let's admit it, even today, not everybody can handle them.
It is if the way to learn is doing it without a tool. Imagine using a robot to lift weights if you want to grow your own muscle mass. "Robot is a tool"
"Growing your own muscle mass" is an artificial goal that exists because of tools. Our bodies evolved under the background assumption that daily back-breaking labor is necessary for survival, and rely on it to stay in good operating conditions. We've since all but eliminated most of that labor for most people - so now we're forced to engage in otherwise pointless activity called "exercise" that's physically hard on purpose, to synthesize physical exertion that no longer happens naturally.
So obviously, your goal is strictly to exert your body, you have to... exert your body. However, if your goal is anything else, then physical effort is not strictly required, and for many people, for many reasons, is often undesirable. Hence machines.
Tool use is fine, when you have the education and experience to use the tools properly, and to troubleshoot and recover when things go wrong.
The use of AI is not just a labour saving device, it allows the user to bypass thinking and learning. It robs the user of an opportunity to grow. If you don't have the experience to know better it may be able to masquerade as a teacher and a problem solver, but beyond a trivial level relying on it is actively harmful to one's education. At some point the user will encounter a problem that has no existing answer in the AI's training dataset, and come to realise they have no real foundation to rely on.
Code generative AI, as it currently exists, is a poisoned chalice.
The point he's making is, we still have to learn to use tools no? There still had to he some knowledge there or else you're just sat sifting through all the crap the AI spits out endlessly for the rest of your life. The op wrote his comment like it's a complete replacement rather than an enhancement.
Tools help us to put layers of abstraction between us and our goals. when things become too abstracted we lose sight of what we're really doing or why. Tools allow us to feel smart and productive while acting stupidly, and against our best interests. So we get fascism and catastrophic climate change, stuff like that. Tools create dependencies. We can't imagine life without our tools.
"We shape our tools and our tools in turn shape us" said Marshall McLuhan.
For learning it can very well be. And also it really depends on the tool and task. Calculator is fine tool. But symbolic solver might be a few steps too far. If you don't already understand the process. And possibly the start and end points.
Problem with AI is that it is often black box tool. And not even deterministic one.
AI as applied today is pretty deterministic. It does get retrained and tuned often in most common applications like ChatGPT, but without any changes, you should expect a deterministic answer.
Copying and pasting from stack overflow is a tool.
Itβs fine to do in some cases, but it certainly gets abused by lazy incurious people.
Tool use in general certainly can be lazy. A car is a tool, but most people would call an able bodied person driving their car to the end of the driveway to get the mail lazy.
mathieuh|1 year ago
Personally I see it more like going to someone who (claims) to know what they're doing and asking them to do it for me. I might be able to watch them at work and maybe get a very general idea of what they're doing but will I actually learn something? I don't think so.
Now, we may point to the fact that previous generations railed at the degeneration of youth through things like pocket calculators or mobile phones but I think there is a massive difference between these things and so-called AI. Where those things were tools obligatorily (if you give a calculator to someone who doesn't know any formulae it will be useless to them), I think so-called AI can just jump straight to giving you the answer.
I personally believe that there are necessary steps that must be passed through to really obtain knowledge and I don't think so-called AI takes you through those steps. I think it will result in a generation of people with markedly fewer and shallower skills than the generations that came before.
necovek|1 year ago
AI will let some people conquer skills otherwise out of their reach, with all the pros and cons of that. It is exactly like the example someone else brought up of not needing to know assembly anymore with higher level languages: true, but those who do know it and can internalize how the machines operate have an easier time when it comes to figuring out the real hard problems and bugs they might hit.
Which means that you only need to learn machine language and assembly superficially, and you have a good chance of being a very good programmer.
However, where I am unsure how the things will unfold is that humans are constantly coming up with different programming languages, frameworks, patterns, because none of the existing ones really fit their mental model or are too much to learn about. Which β to me at least β hints at what I've long claimed: programming is more art than science. With complex interactions between a gazillion of mildly incompatible systems, even more so.
As such, for someone with strong fundamentals, AI tools never provided much of a boon to me (yet). Incidentally, neither did StackOverflow ever help me: I never found a problem that I struggled with that wasn't easily solved with reading the upstream docs or upstream code, and when neither was available or good enough, SO was mostly crickets too.
These days, I rarely do "gruntwork" programming, and only get called in on really hard problems, so the question switches to: how will we train the next generation of software engineers who are going to be called in for those hard problems?
Because let's admit it, even today, not everybody can handle them.
eviks|1 year ago
TeMPOraL|1 year ago
So obviously, your goal is strictly to exert your body, you have to... exert your body. However, if your goal is anything else, then physical effort is not strictly required, and for many people, for many reasons, is often undesirable. Hence machines.
necovek|1 year ago
Your favourite online store is full of devices that'd help there, and they are used in physical therapy too.
logicchains|1 year ago
waste_monk|1 year ago
The use of AI is not just a labour saving device, it allows the user to bypass thinking and learning. It robs the user of an opportunity to grow. If you don't have the experience to know better it may be able to masquerade as a teacher and a problem solver, but beyond a trivial level relying on it is actively harmful to one's education. At some point the user will encounter a problem that has no existing answer in the AI's training dataset, and come to realise they have no real foundation to rely on.
Code generative AI, as it currently exists, is a poisoned chalice.
GlacierFox|1 year ago
4ndr3vv|1 year ago
Matthyze|1 year ago
Fricken|1 year ago
"We shape our tools and our tools in turn shape us" said Marshall McLuhan.
antihipocrat|1 year ago
Ekaros|1 year ago
Problem with AI is that it is often black box tool. And not even deterministic one.
necovek|1 year ago
sarchertech|1 year ago
Itβs fine to do in some cases, but it certainly gets abused by lazy incurious people.
Tool use in general certainly can be lazy. A car is a tool, but most people would call an able bodied person driving their car to the end of the driveway to get the mail lazy.
saagarjha|1 year ago
numpad0|1 year ago
unknown|1 year ago
[deleted]