Forum Moderators: Robert Charlton & goodroi
My robots.txt file is a basic one :
User-agent: *
Disallow:
Sitemap: http://www.example.com/sitemap.xml
Since getting sitemap errors, I modified my sitemap to a basic one containing only 2 URLs :
<?xml version="1.0" encoding="UTF-8"?>
<urlset
xmlns="http://www.google.com/schemas/sitemap/0.84"
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://www.google.com/schemas/sitemap/0.84 [google.com...]<url>
<loc>http://www.example.com/</loc>
<changefreq>daily</changefreq>
<priority>1.00</priority>
</url><url>
<loc>http://www.example.com/ideas.php</loc>
<changefreq>daily</changefreq>
<priority>1.00</priority>
</url></urlset>
As of today, I still get the Network unreachable : robots.txt unreachable error. My web host says that googlebot is crawling other sites. My server logs do not show any errors, so I am at my wits end. I do not know what to check or what to do. Please help anyone?
[edited by: tedster at 5:59 pm (utc) on Aug. 23, 2007]
[edit reason] use example.com [/edit]
I've just tried your suggestion although all the validators I used did not pick up on this. I am assuming what you mean is :
User-agent: *
Disallow:
Sitemap: http://www.example.com/sitemap.xml
implies that the sitemap is out of bounds to all bots because it is not indicated as a separate command whereas :
User-agent: *
Disallow:Sitemap: http://www.example.com/sitemap.xml
Implicitly states that the sitemap URL is to be treated on its own. Am I correct?
But shouldn't such an error return a "syntax error" or "sitempa error" instead of the one I am getting - "Network unreachable - robots.txt unreachable?"