(no title)
dhc02 | 1 year ago
It's taken a while, but the longer we go down this path, the more clear it seems that it is impossible to design a content algorithm that does not have significant negative cultural side effects. This is not to say that content algorithms don't have benefits; they do. It's just that they can't be useful (i.e., designed to optimize for some profitable metric) without causing harm.
I think something like asbestos is a good metaphor: Extremely useful, but the long-term risks outweigh any possible gains.
mandmandam|1 year ago
That's not the pattern I've seen, as close as you are to it.
I've seen lots of platforms be wildly useful. Digg was good for a while; StumpleUpon, Pinterest, Instagram, TikTok, Twitter, Reddit and even Facebook all had periods at the start where they added real value to people's lives.
At some point they start to "optimize for some profitable metric" - and quickly become heinous.
The problem isn't the algorithm; it's that it gets twisted toward profit. And that's basically a tautology - once you start trying to suck money out of the equation for yourself, that juice has to come from somewhere.
I can envision a platform that isn't based on profit being far more useful than harmful - if it can only ward off the manipulations of the yacht class.
S_Bear|1 year ago
unsui|1 year ago
The inevitable enshittification of goods and services once they reach a certain level of maturity (i.e., profitability) basically guarantees that the yachted-classes will be involved.
Given this de-facto inevitability, the original premise (that algorithmic content is eventually a bad thing) makes more sense