(no title)
blksv | 1 year ago
I agree, that AI may be dangerous if used for destructive purposes, or if used for some critical tasks with too much trust (and the hype that "we're so dangerously near a superintelligence" makes the latter much more likely, in my opinion). But that Humanity will be displaced by autonomous server farms? No way.
(As to the original comment, I think that a bunch of nearly(?) demented elders holding nuclear buttons is a much worse (and immediate!) threat than a server farm which we finally conclude to be intelligent.)
0xDEAFBEAD|1 year ago
One gorilla says to the other: "Those human brains are just synapses firing. They depend on nature to survive. Not a problem"
Another way of thinking about it... Suppose we create a server emulation of a highly intelligent, manipulative serial killer, and speed up the emulation so it thinks 1000x as fast as a human. How do you feel about this? Is the fact that it's "just a server farm" reassuring?
blksv|1 year ago
When it comes to AI we have just a lone detached brain, not in control over anything, so that it cannot even "fire" by itself: someone has to provide its inputs.
> Suppose we create a server emulation of a highly intelligent, manipulative serial killer <...> How do you feel about this?
Quite indifferent: the only field I can see for such a simulation is game development, but that would be huge overkill.