Welcome to WebmasterWorld Guest from 220.127.116.11
"When we tested a sample of the URLs from your Sitemap, we found that some URLs were not accessible to Googlebot because they contained too many redirects. Please change the URLs in your Sitemap that redirect and replace them with the destination URL (the redirect target). All valid URLs will still be submitted."
What does this mean and what do I need to do?
I didn't think I had any redirects. It is a small, simple site.
[edited by: BCDesigns at 7:08 pm (utc) on July 30, 2008]
HTTP/1.x 200 OK
The first line in the first section is the url I entered. There are about ten more sections that are things like the style sheet, images, adsense, statcounter, etc.
Is that normal? If so, does that mean that it Google's problem?
Also, how many of the url's in the sitemap should I try?
Try a few of your urls, maybe a dozen or so, and if the results are like what you just saw, then it sounds like googlebot has the problem. Others have also seen a similar problem, and there's nothing you can do about it.
Just one other idea. Install the User Agent Switcher add-on and go to your site with a googlebot user agent. Just in case, somehow or other, your server is treating googlebot differently, this will show it to you. There are even some hackers these days who will get into a server and install a script on your pages that serves special junk just for googlebot.
Sometimes webhosts take some misguided steps, too. So switch out the user-agent and give that a go.
You could also have a look at Googlebot's activity in your server logs to see how and if it gets redirected (likely with a 302 or 301 status code). Log files are pretty much the definitive place to determine spidering activity.
Five messages were cut out into new thread about conflicts
between Live HTTP Headers and an HTML editor. The new thread:
[edited by: tedster at 7:12 pm (utc) on Aug. 1, 2008]