Forum Moderators: DixonJones
I not sure how you would block it if it is a free piece of software?
But because it's a free software - and can be installed on more than 100,000 computers - you can't of course, ban it via all it's IPs, which would seem impossible.
However, you can add the following code to your .htaccess file and forbid access to it. This way I think is the best way to do it, because you can generally block any robot you think is nasty, even if it doesn't obey the Robots.txt Exclusion Standard.
[engelschall.com...]
Blocking of RobotsProblem Description:
How can we block a really annoying robot from retrieving pages of a specific webarea? A /robots.txt file containing entries of the "Robot Exclusion Protocol" is typically not enough to get rid of such a robot.Problem Solution:
We use a ruleset which forbids the URLs of the webarea /~quux/foo/arc/ (perhaps a very deep directory indexed area where the robot traversal would create big server load). We have to make sure that we forbid access only to the particular robot, i.e. just forbidding the host where the robot runs is not enough. This would block users from this host, too. We accomplish this by also matching the User-Agent HTTP header information.RewriteCond %{HTTP_USER_AGENT} ^NameOfBadRobot.*
RewriteCond %{REMOTE_ADDR} ^123\.45\.67\.[8-9]$
RewriteRule ^/~quux/foo/arc/.+ - [F]
Sid