Forum Moderators: goodroi
User-agent: *
Disallow /block/this/
Disallow /blocked/
However when I put in a full address, it approved it:
User-agent: *
Disallow http//www.domain.tld/block/this/
Disallow http//www.domain.tld/blocked/
(left out : so that it doesn't create a true link here, but in my robots.txt it exists)
Anyone else running into this issue?
Didn't know if it really mattered which way is better or should not be used.
It must be in the root directory of the web site as spiders will not look for it anywhere else
To exclude all robots from parts of the server
User-agent: *
Disallow: /cgi-bin/
Disallow: /misc/sitestats/
Exclude a specific spider from parts of the server
User-agent:slurp.so/
Disallow: /cgi-bin/
Disallow: /secure/
Disallow: /products/
Disallow:/misc/sitestats/
This indicates that nothing is disallowed and the spider can follow all links
User-agent: *
Disallow:
To allow a single robot complete access and exclude all others
User-agent: Googlebot/1.0
Disallow:
User-agent: *
Disallow: /
This would prevent your entire web site from being indexed
User-agent: *
Disallow: /
Hope this helps
ncw164x
The site gave me errors when I only tried the directory
/this
/and/this
but when I use a full link, it comes back as 'okay':
http//www.site.tld/this
http//www.site.tld/and/this
I have found that most bots don't listen anyways including Yahoo...they keep going after disallowed.
I'm so disgruntled with the net that I'm going to make my site go stealth.
What the hell is the point for search engines posting how to deal with their bots/spiders if they can't even control them properly?
And for all those worthless sites like DMOZ who fail to update on a timely basis, I have no respect for & could care less about dealing with them...