Forum Moderators: open
I've never seen this before - anyone experience similar requests?
This is a new site, recently submitted. It hasn't yet visited any other pages.
BTW: Here's what my "robots.txt" file looks like. Anyone see an error I'm missing?
User-agent: *
Disallow: /inc/
Disallow: /images/
Disallow: /.status/
Disallow: /reports/
They should look something like this:
Cache-Control: must-revalidate, max-age=7200
Expires: Tue, 05 Aug 2003 21:37:58 GMT
Last-Modified: Sat, 02 Aug 2003 05:40:41 GMT
Jim
Here is the reponse I get testing index.html
and robots.txt, respectively:
HTTP/1.1 200 OK
Date: Tue, 05 Aug 2003 20:44:05 GMT
Server: Apache/1.3.26 (Unix) AuthMySQL/2.20 PHP/4.1.2 mod_gzip/1.3.19.1a mod_ssl/2.8.9 OpenSSL/0.9.6g
X-Powered-By: PHP/4.1.2
Connection: close
Content-Type: text/html
HTTP/1.1 200 OK
Date: Tue, 05 Aug 2003 20:46:14 GMT
Server: Apache/1.3.26 (Unix) AuthMySQL/2.20 PHP/4.1.2 mod_gzip/1.3.19.1a mod_ssl/2.8.9 OpenSSL/0.9.6g
Last-Modified: Tue, 05 Aug 2003 20:23:29 GMT
ETag: "9284f-55-3f301241"
Accept-Ranges: bytes
Content-Length: 85
Connection: close
Content-Type: text/plain
What doesd this mean?
Thanks,
Friday
I find that this one requests robots.txt nearly every time, and there are often several on my site at a time.
> What does this mean?
Sticking to robots.txt, it means that Google is free to apply a default max-age and Expires date to your robots.txt file, because you have not specified otherwise. This is usually not a problem. The problem comes in where the webmaster specifies an unusually short max-age or the Expires date is only a short time in the future. In this case, Google detects that the robots.txt it cached earlier has expired, and so needs to be re-fetched before they proceed with further spidering.
Since you didn't specify these times, you don't need to worry about it unless you change your robots.txt a lot.
Jim