Welcome to WebmasterWorld Guest from

Forum Moderators: goodroi

Message Too Old, No Replies

Robots.txt quick question

4:59 pm on Jan 9, 2011 (gmt 0)

Junior Member

10+ Year Member

joined:Jan 10, 2005
votes: 0


We have a bunch of 404 errors from Webmaster Tools that sit in a directory that has thousands of other valid URLs.

We want to disallow these 404 URLs without affecting the others in the directory like so:

404 error URL: /xyz/keyword/abc/
Valid URL: /xyz/keyword 1/abc/

If I disallow using:

Disallow: /xyz/keyword/abc/

will it disallow the URL above and continue to crawl all other URLs in the /xyz/ or will it try to disallow all of them?

Thanks in advance for your help.
5:12 am on Jan 10, 2011 (gmt 0)

Senior Member from US 

WebmasterWorld Senior Member tangor is a WebmasterWorld Top Contributor of All Time 10+ Year Member Top Contributors Of The Month

joined:Nov 29, 2005
votes: 1050

This sounds like something better handled via .htaccess. Robots.txt is not going to help against those SEs which already found the URIs as they will keep hitting their already collected info to see it the link still exists. Feed it a 410 (gone), 404 (not found, ie, do nothing) or 301 to the page you want the SE to find.

This assumes that the bad URIs are NOT THERE. If they are, then WHY?

Join The Conversation

Moderators and Top Contributors

Hot Threads This Week

Featured Threads

Free SEO Tools

Hire Expert Members