Forum Moderators: open
Slurp is ignoring robots.txt directives. I have blocked a directory in robots.txt 6 months ago. Then I put a spider trap in this directory to trap bad spiders and every few days slurp drops right in it!
Crawling speed is also very quick and causing problems with load on a massive Mysql heavy site. I have used the
User-agent: Slurp
Crawl-delay: 135
Directive for a 135 seconds display, and 1 month on am still getting requests from Slup on the 74.6.x.x ip blocks!
Due to the low volume of traffic I get from yahoo which is only around 200 visitors a day and considering they are the second largest load on my bandwidth I am considering banning them from my server.
I filed a report to Yahoo and so far have had no response and Slurp is still being badly behaved.
I have seen Slurp misbehave before, but not recently. It has never ever fetched my spider-bait, though.
Jim