Forum Moderators: Robert Charlton & goodroi
<META NAME="ROBOTS" CONTENT="NOINDEX, NOFOLLOW">
User-agent: *
Disallow: /product-category-1/
Disallow: /product-category-2/ As the product pages of course take up a huge bulk of the site
Is taking a one or two sentence description that describes the product and turning it into 350+ words that describes all the features and benefits of the product creating adding additional value whilst eliminating duplicate content?
If all the pages use the same header code, you can achieve the same by ...
If I de-indexed the huge amount of problem pages, do you think that by de-indexing a huge chunk of the site would this affect these ones that already rank? De-indexing product pages is a fairly drastic solution right, but I can't really see any other way of doing it because I don't want the site to be viewed as some low-quality scraper thing, which unfortunately it technically is at the moment. I'm trying to fix it though but it seems like it's an insane amount of work!
If I de-indexed the huge amount of problem pages, do you think that by de-indexing a huge chunk of the site would this affect these ones that already rank?
what Small-Medium business would have even considered re-writing a whole catalogue back then
Think of it from the point of view of a search engine looking to produce as diverse a set of results as possible
I wouldn't trust the search engine to make this decision on its own. That's what the "canonical" tag is for.
the chances of being penalized as a scraper are practically non-existent
I do find it odd though, all the hoopla about duplicate content within a single site/domain.
I believe the OP is talking about duplicate content outside a single domain, and thin content to boot. It's not a question of reducing the internal duplicate content, it's an attempt to jump start the balance between good content and low quality content.