Forum Moderators: Robert Charlton & goodroi
If you did, they would know about this file and spider it. However as of date, no bot can interprete this file and crawl the included URLs. Even Google bot does not.
You have to submit your file to the Google Sitemaps page and a special software interpretes this file and prepares a list of files to crawl, that is given to the Gbot.
Case in point. This is stated on the protocol pages. You might even find more ambiguities...
Please note that the Sitemap Protocol supplements, but does not replace, the crawl-based mechanisms that search engines already use to discover URLs. By submitting a Sitemap (or Sitemaps) to a search engine, you will help that engine's crawlers to do a better job of crawling your site.
mmm... Sounds like they are referring to more search engines then just Google.
And then this...
The Sitemap Protocol allows you to inform search engines about URLs on your websites that are available for crawling. In its simplest form, a Sitemap that uses the Google Sitemap Protocol is an XML file that lists URLs for a site.
Once again, it sound like they are talking of multiple search engines.
And this, on the FAQ page...
Your Sitemap should be placed in a location on your web server that is accessible to our crawlers. If you'd like to prevent other robots from accessing your Sitemap file, you may use a robots.txt file that complies with the Robots Exclusion Standard.
Other robots? Are they implying that other bots can crawl this page?
Good question!