Forum Moderators: Robert Charlton & goodroi
As far as downsides, most of the troubles I've heard of come from technically inaccurate sitemap generators. You want to be sure that its not including diferent urls that resolve to the same content. This has particularly been a challenge for some sites that are using a rewrite scheme and only want to see the "friendly" urls in the index - but the generator also picks up the "unfriendly" urls that use a query string.
One thing that I need to mention is that query string URL's should never be available to anyone if you are using mod rewrite ! If a crawler tool can find so can google !
[edited by: tedster at 6:00 pm (utc) on May 22, 2008]
I'd trust the site: operator results over Webmaster Tools, which unfortunately tends to give buggy numbers more frequently. After all, site: gives actual proof that the url is indexed, whereas Webmaster Tools is a secondary reporting utility. More steps involved means more potential points of failure.
Also, it's very unlikely that Google will actually index every url in sitemap.xml file. Spider them all? Maybe, but spidering is no guarantee that a url will stick in the index. However, you should be able to hit 80% - 90% indexed with a strong website, and it sounds like you're in that area.
[edited by: tedster at 4:04 am (utc) on May 26, 2008]
However, you should be able to hit 80% - 90% indexed with a strong website
The only comment from Google I have found is from the Help Center where they state, "While we can't guarantee that all pages of a site will consistently appear in our index, we do offer our guidelines for maintaining a Google-friendly site".