(no title)
harvie
|
6 months ago
Maybe we can just configure webservers to block anyone who requests robots.txt, regular browsers don't do it, but robots do to get list of urls to crawl (while ignoring rules). Just create simple PHP/CGI script that adds client IP addres to iptables once /robots.txt is accessed.
Trung0246|6 months ago
robots txt file size is usually small and would not alert external services.