I have only seen 'Sitemap:' (with a capital 'S') - although I'm not sure whether it is case-sensitive or not? (All other directives have a capital first letter)
I would also put the Sitemap: directive last (and separated by a blank line - used to delimit records in robots.txt). I have read that not all robots support the Sitemap: directive, so in order to prevent these bots from prematurely aborting processing of the robots.txt file it should appear last.
The 2 pages I want to exclude are included in the sitemap (it gets automatically generated by my cms)
IMHO, if they are disallowed in robots.txt then the search engine should be prevented from accessing them, regardless of whether they are linked to from elsewhere, or included in your sitemap - but I don't know for sure; just my opinion. Ideally they should not be in your sitemap.
Msg#: 4084471 posted 8:47 am on Apr 9, 2010 (gmt 0)
by default the search engine bots will crawl your site unless excluded by some technical method. a missing or empty robots.txt file is equivalent to permission to crawl. an empty robots.txt is preferable since the frequent requests for that file will return a 200 OK status code response instead if a 404 Not Found. if you wish to exclude some or all bots from crawling a part of your site, the robots.txt file is one of the methods available.