Forum Moderators: Robert Charlton & goodroi

Message Too Old, No Replies

Duplicate content - what proportion is dangerous to your site ?

         

Whitey

3:33 am on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Are small proportions of duplicate content safe? By small I mean 5 or 10 %.

When does it start to drag a site off of the SERP's ?

tedster

2:43 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Are you talking about duplicate content within one domain - or content that is duplicated on another domain?

ichthyous

3:10 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



This is also something I have been wondering about. In my case it's all on the same domain. Various pages which WMT is reporting as having dupe titles and descriptions. Are dupe titles a more serious problem than descriptions?

bwnbwn

3:51 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



WMT is reporting as having dupe titles and descriptions. Are dupe titles a more serious problem than descriptions?
this is not related to content at all. Google is teling you there is a problem here and all the titles and description are the same. Title and descriptions tell the bot what the page is about in very short sentence. If all titles are the same it is telling the bot all the content is the same, then yes you got a serious problem.

Most of the time this leads to the pages being put in supplemental or deindexed and you won't get squat for searches from all the content on the pages. Titles and descriptions have to tell the bot what the page content is related to.

Whitey
Really nobody can tell you the answer on the % of content that will get a site hit by a filter. This is a subject that is different for ecommerce sites as it is for informational site.

Ecommerce sites is a different world as there may be 1000 sites using the same product description on the net with 1000's of pages pretty much the same. Kind of hard to name a widget different 1000 times, so you question is a tough one to answer.

Best possible answer is if you have only 5% why not take the time and make them different then you have 0% and can rest assured your ok.

Whitey

9:39 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Are you talking about duplicate content within one domain - or content that is duplicated on another domain?

I guess i was thinking both, plus including meta titles and descriptions in that basket.

JoeSinkwitz

9:43 pm on Aug 7, 2008 (gmt 0)

10+ Year Member



They've been really sensitive to on-domain duplication of metas and titles lately...the % threshold appears to be have been tightened on 8/1.

Quadrille

9:49 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



duplication of <title>, meta description and / or content, is taken very seriously by SEs, especially Google;

There's no fixed percentage - but they will take into account the whole page - so a page with 44 links, a few promo sentences, and other stuff common to every page - and just a para or two of 'unique' content, may be in Big Trouble.

tedster

10:32 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I agree that this area seems to be tighter since 8/1 - and I wonder why. I'm wondering if Google has deployed another database partition (similar to the Supplemental Index) and those top-level tagging areas are now even more important for Google to retrieve results from a grwoing assortment of partitions.

Whitey

10:52 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



.... hmmm .

Everything on Google is mathematical and there has to be a "tipping point" [en.wikipedia.org] - "the levels at which the momentum for change becomes unstoppable."

We know that if you have sitewide duplication though canonical problems a site will disappear from the results. Maybe this is because the duplicated pages could/may be scored at minus levels [ or at least zero ], sucking the entire site backwards, and out of the results. This type of thing is black and white where 100% of the site is affected.

But what about those "occassional" duplicate meta tags and titles in WMT that you may be looking at which have slipped under your radar?

What about the occassional interlinking between same content on different domains?

What about the %'s of snippet content between same site URL's or from other domains to help you build content, albeit not unique.

Complex probably. So let's break it down. What about if everything else on your site was perfect except meta descriptions and meta titles. What % of this would you consider critical to this tipping point ?

[Sorry Tedster ] - our posts overlapped

[edited by: Whitey at 11:02 pm (utc) on Aug. 7, 2008]

Whitey

11:04 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



I'm wondering if Google has deployed another database partition (similar to the Supplemental Index)

or several ?

Quadrille

11:49 pm on Aug 7, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Complex probably. So let's break it down. What about if everything else on your site was perfect except meta descriptions and meta titles. What % of this would you consider critical to this tipping point ?

I don't think the 'tipping point' exists in isolation, though I totally agree there'll be a mathematical formula in there somewhere.

We know the 'amount of duplicated content' matters; this is probably a proportion of the code, rather than a fixed amount.

But Google may also recognise 'blocks' of duplicated content, rather than considering one total value.

Google may vary the requirement with the total page size.

Google may take duplicate content more seriously if the <title> is also duplicated.

Google may be more concerned about duplication if duplicated links are involved.

And there are many other relevant factors that 'may' be involved.

I very much doubt it's a simple % - and if it was, we'd have had serious estimates to argue about by now!

[edited by: Quadrille at 11:51 pm (utc) on Aug. 7, 2008]