Forum Moderators: goodroi
For instance, my real homepage is located here on the server: /www/index.asp
And my development file is here: /dev/index.asp
I'm wondering if each directory needs its own robots.txt file? Or is it as simple as this:
# Google
User-agent: googlebot
Disallow: /dev/
Thanks for your help in advance...
Joe Bray
User-agent: *
Disallow: /dev/
this page has the robots.txt standard [robotstxt.org].
I have it inside the production website directory;
/www/robots.txt
But if I'm understanding you correctly, it should be located here instead:
/robots.txt
Does that sound right? Is there any way to test this sort of thing?
I put the modified robots.txt file at the root level: /robots.txt
And I also left the old one where it was: /www/robots.txt
What I will do is check back tomorrow and look in the Google Webmaster Tools, and see which robots.txt Google has cached for the website. Hopefully I will see the modified one, so that I can delete the other.
Joe
So, what I need to do is create a second robots.txt and place it into the other directory - the root of the development page.
Thanks for helping me work thru this...
Joe
So, what I need to do is create a second robots.txt and place it into the other directory - the root of the development page.
therefore use the following files...
/www/robots.txt:
User-agent: *
Disallow:
/dev/robots.txt:
User-agent: *
Disallow: /
you can use the robots.txt tool in the google webmaster tools to verify which urls are allowed and disallowed by googlebot.
you can make tweaks to code from the cached version in the form and then update the file on your site with the final version.
not sure how often they update cache with a new file...