Forum Moderators: Robert Charlton & goodroi
The site seems to have been in a holding pattern for the last few months as far as Google traffic is concerned.
Interestingly, the site does well for long tail search terms and extremely well for competitive image queries (by competitive...3,000+ unique visits per day from a single query). The site isn't optimized for images and has relatively few images in proportion to the rest of the content. New content is added to the blog daily.
I'm confident that the original hack was resolved and there has been no other wordpress corruptions since the original incident. But I don't know whether a site:example.com 'drugname' query a month, 2 months, or 3 months ago was returning the cached hack results. I imagine it's more likely that the query was returning this same result in the interim. Seem more likely than six month old cache results returning.
Any ideas?
[edited by: tedster at 12:07 am (utc) on Feb. 19, 2009]
[edit reason] switch to example.com - it cannot be owned [/edit]
Tedster was correct. The hack was cloaking for the Googlebot user agent. The code was removed in late February.
But the SERPs are still showing the same cached results for site:example.com "spam drug name".
I cannot see any of the spam URLs that Google has cached for a page when I view that page as Googlebot. Any ideas?
It's possible to cloak via IP, which you will not be able to detect with a browser. If the cache post-dates your fix for the hack, I would check for the hack again - looking particularly at mechanisms for IP delivery - server config (including htaccess files) and any dynamically-generated content.
Whether the site has suffered at all is something you can only diagnose by its performance. If the URLs have an "expected" rank then you don't have any problem.
what is a reasonable amount of time in which Google will update the cache?
Never. 6 months. 2 weeks. 10 minutes ;)
It depends on the links directly to the URL, and whether Google's spidering algorithms believe it to be content worthy of checking on frequently (mostly the former if you ask me).
IMO, a telling factor is whether googlebot requests the URL, but that copy is not retrievable in results.
I had a feeling that would be the answer
It's the range of possible answers ;)
You can look at an individual URL and get a reasonable idea of how frequently you would expect it to be cached - primarily based on past spidering activity and the "strength" of links to the page. Otherwise, your guess is as good as mine :)