Forum Moderators: open
1 bot is doing 99Gb per month worth of requests.If they're disregarding a robots.txt Disallow, then yeah, you'll have to block 'em. (I am not prepared to believe that any robot in the world, including bingbot, eats 99GB worth of robots.txt in a month.)
RewriteCond %{HTTP_USER_AGENT} (Ahrefs|Access|appid|Blex) [NC,OR]
RewriteCond %{HTTP_USER_AGENT} (Capture|Client|Copy|crawl|curl) [NC,OR]
RewriteCond %{HTTP_USER_AGENT} (Data|devSoft|Domain|download) [NC,OR]
RewriteCond %{HTTP_USER_AGENT} (Engine|fetch|filter|genieo) [NC,OR]
RewriteCond %{HTTP_USER_AGENT} (Jakarta|Java|Library|link|libww) [NC,OR]
RewriteCond %{HTTP_USER_AGENT} (MJ12bot|nutch|Preview|Proxy|Publish) [NC,OR]
RewriteCond %{HTTP_USER_AGENT} (scraper|Semrush|spider) [NC,OR]
RewriteCond %{HTTP_USER_AGENT} (Wget|Win32|WinHttp) [NC]
RewriteRule .* - [F] The bots are not actually hitting the server, they are hitting the websites on it.An interesting way of putting it, but technically correct. What your hosts are essentially saying is: It's not our problem, it's yours. Some sites may welcome visits from some of the named robots, so it would be wildly inappropriate for the entire server to deny access.
99% was from Blexbot downloading my photos.How very odd--both because in my experience they're robots.txt compliant, and because I’ve never seen them request an image.