(no title)
rektide | 2 years ago
Now, I personally I think it's trying to swim up the waterfall & ultimately worse for everyone, but: Mastadon specifically has had a strong history of being anti-search, anti-scraping. You aren't supposed to be surveiling folks at industrial scale on the fediverse.
There's widespread skepticism about Meta respecting rules of the road. Having a huge giant shark join the pool of lots of little fish seems like a scary proposition. How we can still protect & have sovereignty over our different fedi-sites is a real question when there's a company with so much technical, economic, and popular leverages.
dahwolf|2 years ago
I'm honestly pretty skeptical about the fediverse aspect of Threads. It suggests that if I open a new fediverse instance and follow their accounts, I can suck in their timeline and do with it whatever I want. In particular, to bypass ads.
Hence, I could make a "best of Threads" fediverse instance without ads. Or maybe put my own ads on it.
Or, I could build my own client on top of the Threads instance.
None of this sounds very Meta to me.
WorldMaker|2 years ago
It's the opposite of naive: it's extremely well thought out and heavily deliberated. Making so many things "public" by default is an invite to people. It's an intentional welcome mat in old school "Internet 1.0" sort of way. But just because you want to welcome people doesn't mean you have to welcome robots (crawlers, etc). Many instances do that deliberately in a very old school "Internet 1.0" way by saying so in their ROBOTS.TXT file (in addition to other places).
In the old web, crawlers were expected to read ROBOTS.TXT and no matter how "public" they thought the website was they found, ROBOTS.TXT was supposed to be the final word.
Anyone scraping or making searchable "at will" random chunks of the Fediverse is easily violating some number of ROBOTS.TXT files. That is an ancient technical convention that isn't new or naive. The internet knew even then that bad actors would ignore ROBOTS.TXT files. The old internet learned to name and shame the bad actors, and in some cases would back that up by force with firewall blocks and in some cases lawsuits. Mastodon does that too. That's why a lot of Mastodon instances are preemptively blocking Threads, because they don't trust Meta to follow good behaviors such as checking ROBOTS.TXT, because Meta hasn't shown a history of being a good actor there and because Thread's privacy policies seem to imply that they don't care to be a good actor for their own users (to the point of not supporting EU users at all because GDPR is "too hard"), so it makes it much harder to assume they will be good actors with respect to all of the conventions around Mastodon data including the classic ROBOTS.TXT.
The Mastodon culture of "public for people, but not for ROBOTS, or only select ROBOTS" is an ancient internet tradition. It's hard to call that naive, when it has decades of history and internet social norms (including good outcomes) behind it. What's naive is thinking that because some major corporations stopped respecting good social norms in the name of increased ad revenue that those norms no longer apply and "anything technically possible is allowable". Read the ROBOTS.TXT in the room and stop being motivated by technology for technology's sake without respecting ethics. Be a good actor in any ecosystem.