Forum Moderators: DixonJones
this bot leaves no User Agent and nothing else an example :
[01/Feb/2002:15:30:17 +0000] "GET / HTTP/1.0" 200 20253 "-" "-"
I can only use robots.txt, will this work
User-Agent:
Disallow: /
or
User-Agent: -
Disallow: /
Any ideas, this bot is really starting to get on my nerves.
Thanks in advance as always
Dazz
User-agent: *
Disallow: /
(note the wildcard "*")
[robotstxt.org...]
As apparently everybody else, I'd be surprised if a robot who's maintainer has neglected to configure an UA string would bother to read robots.txt anyway. After all, the same informal standard that defines robots.txt also requires that each robot identify itself.
There is a problem with banning a visitor because the browser doesn't contain a user agent. E.g. with some proxy services, you may inadvertently ban an visitor when the browser requests a javascript or css file. These types of request are often done through a different IP that has no user agent.
Key_Master, you could work around that by putting your javascript and css files in a different a sub-folder with a .htaccess that does not have the above restriction
Or:
SetEnvIf User-Agent ^$ keep_out
<Files ~ "(\.html¦\.jpg¦\.gif¦\.what_ever_else)$">
order allow,deny
allow from all
deny from env=keep_out
</Files>
>The looksmart spider doesn't use one either...
It won't matter in six months when they are out of business..
[webmasterworld.com...]
Now you know they have one that doesn't use an agent. What's it for? I have no idea. Perhaps it checks for cloaked sites or possibly even ad listings on free submit sites.
Looks like they dumped the UA string some time between August and November. Since it only fetches URLs that are listed in the directory, I assume it is still the same link checker than before. Another interesting deatil I noted is that they switched from HEAD to GET requests between July and August last year.
I serve empty pages (without an error) to visitors without an UA, so far without any negative side effects on my LookSmart listings. Guess I'll start making an exception for them, just to be on the safe side.