Forum Moderators: martinibuster
Thanks.
I'm afraid there is no solution for this. You need to have the robots.txt to shut off certain areas of your site where spiders are not supposed to be.
But at the same time, this is - of course - a dead giveaway for anybody who means harm. I myself read robots.txt of competitors sites regularly, and any cracker/hacker who's looking for interesting stuff is bound to look at it too.
There are so called spambot-traps, which use robots.txt to lure harmful bots into visiting non-permitted pages so they can lock out the IP-address. But it's only a partial solution.
so if i had, e.g.:
User-agent: *
Disallow: /private/
then what a cracker would have to do is search from the root and follow links manually (unless programs like xenu's link sleuth disregard 'robots.txt/meta noindex nofollow' info -- i'll have to test this out if nobody already has) until s/he found their way inside.