Interesting point of view, didn't know about Jevon's paradox before. To me, the outcome still depends on whether AI can get superhuman [1] (and beyond) at some point. If it can, then, well, we will likely indeed see that suitable-for-human areas of the intellectual labor are shrinking. If it cannot, then it becomes an even more philosophical question similar to the agnosticism beliefs. Is the universe completely knowable? Because if it's not, then we might as well have an infinite more hard problems, and AI just rises a bar for what we can achieve by paring a human with AI compared to just human alone.[1] I know it's a bit hard to define, but I'd vaguely say that it's significantly better in the majority of intelligence areas than the vast majority of the population. Also it should be scalable. If we can make it slightly better than human by burning the entire Earth's energy, then it doesn't make much sense.
No comments yet.