Forum Moderators: Robert Charlton & goodroi
When I tested this URL against our Robots.txt in G's Webmaster Tools, it showed up as BLOCKED, yet it still shows up in SERPs.
In the meantime, we lost almost all even-remotely longtail traffic virtually overnight. I'm not sure yet if the two are related, but wanted to reach out and see if anyone else has seen anything similar.
UPDATE: Now hundreds of URLs in Disallowed directory showing up in Serps.
albeit without title or snippet
That happens when backlinks give Google enough information about a url to show it in specific search results.
Robots.txt has prevented actual spidering, but these url-only results still may appear. If it's important not to show that result, use the url-removal tool in Webmaster Tools.
I'm pretty sure that it's not related to traffic loss.
Still, the same steps would be needed - make sure the robots.txt does what you intend and then use the url removal tool. Or, you could make sure everything in the blocked directory has a robots meta tag that says "noindex".
URLs which shouldn't be included and which I had removed via the GWT tool a few months ago started reappearing...
And they are duplicates of the international version of the site so suspect is the cause of the traffic drop (about 20% drop than the average last week).
robots.txt tells Google not to spider the page. The meta robots tag (in the following format) tells Google not to index the page or any references to it....
<meta name="robots" content="noindex">
If you use both robots.txt and the robots meta, Google won't spider the page and thus won't see the robots meta and won't know that the page shouldn't be indexed. In such a situation, if there is a reference (ie, link) to the page on an unblocked page, Google might index the reference to the page and return the url.
[edited by: Robert_Charlton at 8:43 pm (utc) on June 8, 2009]