top | item 44683989

(no title)

jaydenmilne | 7 months ago

ArchiveTeam is trying to brute force the entire URL space before its too late. You can run a Virtualbox VM/docker image (ArchiveTeam Warrior) to help (unique IPs are needed). I've been running it for a couple months and found a million.

https://wiki.archiveteam.org/index.php/ArchiveTeam_Warrior

discuss

order

localtoast|7 months ago

Docker container FTW. Thanks for the heads-up - this is a project I will happily throw a Hetzner server at.

chneu|7 months ago

im about to go setup my spare n100 just for this project. If all it uses is a lil bandwidth then that's perfect for my 10gbps fiber and n100.

wobfan|7 months ago

Same here. I am geniunely asking myself for what though. I mean, they'll receive a list of the linked domains, but what will they do with that?

hadrien01|7 months ago

After a while I started to get "Google asks for a login" errors. Should I just keep going? There's no indication on what I should do on the ArchiveTeam wiki

ojo-rojo|7 months ago

Thanks for sharing this. I've often felt that the ease by which we can erase digital content makes our time period susceptible to a digital dark ages to archaeologists studying history a few thousand years from now.

Us preserving digital archives is a good step. I guess making hard copies would be the next step.

AstroBen|7 months ago

Just started, super easy to set up

cedws|7 months ago

Why wouldn’t Google just publish a database of URLs? Even just a CSV file? Infuriating.

devrandoom|7 months ago

I suspect there are links to some really bad shit in there. Google is probably in damage control mode.