Forum Moderators: open
If someone has stolen your content, you could be penalized with the duplicate content, even though you are the rightful owner, and yours appeared first. In theory, Google should discount the newest duplicate page. But this is not what I have seen in practice since the new duplicate filter was introduced a couple of months ago.
Another possibility is if some of your content is taken from another site - ie. product description pages etc. This could also trigger the duplicate content filter.
Page 1 - Guide to Red Widget Shops
Page 2 - Guide to Blue Widget Shops
If you have one shop that falls into both categories and you carry the same description on each page is this dodgy.
Or are we talking about straightforward copying other peoples content?
You do get penalized within your own site, but it doesn't matter. All that happens is Google nominates one page as good and the rest duplicates. This means users still find your good page.
If you have similar content I haven't seen that penalized.
[edited by: SlyOldDog at 10:35 pm (utc) on Sep. 7, 2003]
I've just done a search and about 20% of our pages are deemed duplicate by Google. That seems high and I'm pretty sure we don't have any on our site, although most of our pages are similar because we have the same content in several languages.
So maybe you do get good content dropped after all.
I then plug it into Google with " " around it and then click the filter link to bring up all the results, and see what you find. The results will often surprise you.
I am not aware of a tool to do this automatically. Perhaps someone has done something with the Google API, because automated queries are not permitted otherwise.
This is close to being duplicate content (apart from their own header they display above my page).
I don't really know what these sites are for, but they must be creating a lot of duplicate content.
It's a jungle out there!
I've just done a search and about 20% of our pages are deemed duplicate by Google.
What is the best way to determin this? Is it just allinurl:www.domain.com? When I do this only 2 URLs show, and then the old
"In order to show you the most relevant results, we have omitted some entries very similar to the 2 already displayed."
Surely this doesn't mean Google considers almost all of my site to be duplicate?
or should I just do a
site:www.domain.com keyword
to determin what pages if any are considered duplicates by Google for that keyword?
Thanks
Just do the allinurl:mydomain.com search and check how many results get returned (the number on the top right corner of the screen.
Then go to the navigation bar and add &filter=0 onto the end of the existing query.
Now you may get a different number of results returned. The difference between the first number and the second is pages Google thinks are duplicate.