Forum Moderators: Robert Charlton & goodroi
It's harder for copied content to beat a strong website - but from time to time it happens. During some periods of adjustment in Google, it can get pretty rough.
If you close one website down which for some reason used your works or whatever you call those articles which were made and usually are a derivative from other peoples works anyway - call it "inspiration" ... five other websites has somehow managed to also take some of your content and republish it and all in the time it took to call your lawyer and begin the first case, then rinse and repeat - while the first domain magically reappears on another server or on another domain under another name -
For me, and that's just me - chasing these ghosts are a waste of my limited resources, so I simply choose to ignore it and focus on what I am actually doing, not what everybody else is - in the end - i'll still die and nobody will remember me anyways.
You only have the advantage of being the first if you create something - and the rest is simply competition who either wants you out of business, or a piece of the cake.. it's the free and unregulated market in effect - and it's a total mess on a global platform like this, too many interests in play with different regulations of how, where, when, who, why and just forget it.
The thing with the internet and the undernet as it is right now, and how it works: Either you adapt to the conditions of it, or you find another media/market/platform which suits you better and are better regulated towards your interests -
It's not going to get updated ever and they are not going to like it if it requires some work to just clean it up so it's not a huge advertisement for you - anyways, also expect the scrapers that they will leave the contents to rot, they don't care about it. Their weakness, is your strength - along with some common sense.
When you write an article, always include your domain name a few times in the text both in a visible color, and with the background color as in www.domain.com a few smart places in the article ... remember, the large scale scrapers are lazy - they are not going to remove it... the same goes with a headline, put your domain name there - just make it take a little work to actually "fix" your contents .... and often they will just go somewhere else because it's not easy anymore...
Also, don't go for anybody and everybody who takes your contents and throw it around - just go for those who is beating you at your own game. Always target your effort on the major problem, not the guy who just thought "this was so cool" i'll take this article it's mine mine and clean it up and miiiiiiiine . (you know, that little guy gollum - he's not the problem, Sauron is)
It's a game, if you're a good player, you'll figure out more creative ways to make it a really sad story to republish your contents on a large scale - by working on internal factors in your own site, not trying to take everybody elses down. I repeat: People are lazy. You aren't - figure out who the best player is ;)
Second, if you do re-write, don't abandon your original content. There's obviously a market for it, so arrange for it to be used on other sites, by agreement and with appropriate links.
Since there are so many people copying pages nowadays, you would have to hire someone part to full time to keep up with the mess of contacting for removal.
Like Ted says, the strong site almost always wins out when Google is stable and running correct'y :P
i can search for various snippets from tripadvisor customer reviews that i found being used on another website and tripadvsior only shows if u click for the omitted results...now id have said tripadvisor was so heavyweight it is always going to rank for text that appeared on its site first....so its quite puzzling as to what criteria is really being used for identical content..
i can search for various snippets from tripadvisor customer reviews that i found being used on another website and tripadvsior only shows if u click for the omitted results...now id have said tripadvisor was so heavyweight it is always going to rank for text that appeared on its site first....so its quite puzzling as to what criteria is really being used for identical content..
I had wondered about that since May and got a partial answer to this with some of the DC’s today. My site, which had never disappeared from Google, did. In its place were dozens of sites that had duplicated my content. This was very interesting in that I’m an expert in finding duplicate content but this junk surfaced out of the "middle of nowhere". It doesn’t appear in MSN or Yahoo and never did. Even more interesting is these sites utilized every conceivable dirty trick you can think of. A quick view revealed at least ten running site maps of my site and replacing my url’s with theirs. Also cloakers of every type, scuzz sites, previously unseen scrappers, and forums that posted entire stolen pages were to numerous to count.
Bottom line is Universal search is crawling the underbelly of the Internet and letting every huckster and schemer around into the results. Duplicate content, which might not have affected your site previously, may well do so in the future. Google’s better idea (ROFL) may soon increase a webmaster’s work even more in combating duplicate content.
i understand the thoughts of many about strengthening your site but it still leaves the question open to my mind of when faced with sites with multiple identical content why so often its the original site that gets buried....without that answer you have no choice but to spend more and more of your time simply dealing with scrapers to reduce the identical copy out there...and to be honest so often the factor here is no about the strength of your site imho....so many low quality sites are riding on the back of scraped content...theres an unknown factor here somewhere..
some have said certain filters are only play for the top ranked sites..so while top sites can get binned into nowhere land it leaves the muck floating around the 40ish plus places...its one possible explanation..though it doesnt explain why the sites at the top get binned for their own content...you may say its not because of that but they probably get binned for other issues...yet so often when the scrapers are taken down those pages re-appear..
scuzz sites, previously unseen scrappers, and forums that posted entire stolen pages were to numerous to count
I always held Google in high regard, but after comparing Google's handling of these trash sites with Y! and MSN, I'd say Google is being completely overrun by scrapers and auto-gen'd trash.
Having such garbage in your index is one thing, but RANKING the trash pages number one for snippet searches while filtering out the CONTENT ORIGINATOR behind the "repeat search" function is truly laughable (for everyone except the people producing the content).
I find it very ironic that Google's "anti-spam" team is so touted within the industry, yet blackhat seo sites/tools are openly RECOMMENDING the use of Google Blog Search and Google News as scraping tools for both downloading the scraped content AND uploading the link-injected garbage.
But, what's most ironic of all is that I have to spend so much time sending DMCA notices to Google when at Y! and MSN the garbage pages are handled correctly by their algos.
For a search on an exact quote, the scraper (a site subsidized by Adsense) will turn up, you have to click to see more results to see mine. This isn't uncommon at all, I'm seeing it a lot.
Cutts was arguing sarcastically, in one of his posts, “what do you expect for free” but the last time I looked Adwords wasn’t free. Perhaps if Google incorporated some type of fee or expulsion for scraper Adsense sites then some of this copying mess might be choked off. There’s little threat of that though. Google profits too much from this mess.