Forum Moderators: phranque
SetEnvIfNoCase User-Agent "^WebZIP" bad_bot
..... (A very long list here)
SetEnvIfNoCase User-Agent "^e-collector" bad_bot
deny from env=bad_bot
However, I am afraid this will slow down the opening speed of all web pages. Which method are you using to block these bad agents? (The robots.txt is useless, IMO)
Yes, it does slow them down, but you might want to consider the other delays in the system, such as the transmission delay from server to client, server disk access latency, etc. These other delays are measured in milliseconds rather than microseconds, and will probably swamp any delay due to native Apache directive execution.
If your site gets fewer that, say, 7200 unique visitors per hour, I doubt you'll notice *any* additional delay if you use fewer than 200 additonal access-control directives.
I've implemented access controls using both mod_access and mod_rewrite directives. As you point out, robots.txt must be recognized as a request to properly-implemented, non-malicious automated user-agents. It does no good if the robot is not coded properly or if it is malicious.
Jim