mhansen - 6:22 pm on Aug 5, 2010 (gmt 0)
I've speculated elsewhere (including our Updates thread [webmasterworld.com]) that since the Mayday update, Google is getting original source attribution wrong more often than they were, ranking the scraped or mashed-up URL and filtering the original. My theory? It comes from Mayday giving good rankings to "sites" they feel are more popular - and therefore better over all destinations for the search user. The emphasis used to be more on the "page" rather than the "site".
After a rather stressful morning with finding my own content outranking me on scraper sites... I thought I would share that in my case, scraper sites outranked my own for 45minutes - 3hours, due to site caching. (content cached, rss was not)
Immediately after I published a new article (WordPress) I started seeing the content show on scraper sites. I spotted the updated sitemap.xml, I saw my new content in feedburner. I could not however find it on Google.
It was 60 minutes or so later that I was able to find it on Google, well... content that led to my domain, and not scrapers.
For myself, it was purely cache related... but made me wonder if Gorgle now treats the scraper site as the original publisher!?!
Several hours later (now) I find my own content leading the serp's, but it took 3-4 hours to wash out. If someone like a ehow or answers were to grab my rss feed and do the same, I would be screwed.