Forum Moderators: open
I have several sites with seperate domain names and servers. Some of those sites are similar in content and the others are more in depth versions of the content.
As an example, lets say the similar sites are a directory of Animals divided into subcategories like Cats, Dogs, etc. Each of these similar sites has it's own layout and design, but the main portion of content, the directory listings with descriptions, are basically the same. In addition, the titles and descriptions of each site focus on keyword synonyms (one site will be titled Dogs, the similar page on the second site will be Canines, etc) to cover the different ways people may search.
Is creating similar sites for synonym keywords against any of the Google guidelines? Is this considered substantially duplicate content?
These sites were created over a year ago and have been happily listed ever since. The're well designed and useful for whoever finds them. However, each Google update makes me nervous that they'll be delisted.
A good rule of thumb is - are you doing this for the users or the search engines.
If the answer is the latter - you can't very well expect that google will like it.
I wouldn't worry about getting banned - unless you are being abusive. Most likely if google's filter picks it up - they will drop one of the pages.
Ton of people and sites have duplicate content. There is not necessarily anything wrong with this. Google just trys not to show more than one copy.
If you type in a few words from "The Raven" for example - they just don't give you 10 million copies of the raven.
compare:
[google.com...]
with:
[search.lycos.com...]
If your title is different and a decent part of the content is different - you should be ok. Layout can be the same - most sites have the same layout.
Google isn't going to let you know the exact amount - as it is a secret. Just keep the principles in mind and you should be ok.
I've got the same problems here, not synonyms but in other languages. I've translated these pages "wholesale" and Google seems to think they are duplicates.
Any way around this?
Thanks and a happy and successful 2003 to all!
I know where you're coming from. I have two sites that have a few pages that are basically duplicates and I eventually want to remove the ones from the original site for fear of being penalized. Unfortunately that site is the one that is getting 80% of the traffic as the other is only two months old. Once the new catches up with the old I plan a complete makeover.
my3cents...I like that! I've used my .06 many times and mentioned inflation as a reason, but nobdy notices. Good originality!
[edited by: my3cents at 12:04 am (utc) on Jan. 4, 2003]
reading this thread worries me, should we consider this duplicate content? and if so should we put a robots noindex tag on all incidences of duplicate content?
Hope this info helps.
Using subdomains or seperate urls will help you short term because you can get several listings on one google page and you can boost your pagerank artificially, but you may get a slap on the wrist for it too if you get reported.
There's no problem optimising one site for all the keywords you please. Just make a seperate page for each set of keywords.
We only have 2 domains per subject area, and the second one is disaster backup only in case the main site gets dropped. We don't promote the second site and it normally doesn't make the page 1 SERPS. If we get dropped, we'll ask the backward linking sites to move their links to the new site. This really is the only way to sleep at night. :)
It seems to me that what I did is what doorway pages were originally used for... basically to direct people to a relevant page when they're searching using different keywords. Google doesn't like doorway pages because "Google does not encourage the use of doorway pages. We want to point users to content pages, not to doorways or splash screens." Instead, I have skipped the doorway page and created a page of content where a doorway page would have been. Isn't this a good thing? The intent is not to fool a search engine or a searcher, it's merely to get my content to those who are looking for it in different ways.
I think it's common sense people, if you're trying to cheat the search engines, and you are spreading content across multiple domain for the sake of getting better rankings and targeting more phrases, it is going to look like spam or at least put you at risk of being reported. I have had very good luck just building lots of content on single site and having my internal pages outrank competitors for common phrases. It really does work and the larger your website is and the more pages of content it has, the better internal PR you have. All of these factors help boost your main page to the top on several variations.
however this is not how it has worked in practice. we tried the one big site approach, however this has not been as successful as targetting niches with a little bit of content crossover.
this is why we have multiple domains. it has been done for the customers best interests, we do not crosslink either. surely this is not a bannable offence?
There are thousands of sites using the same ODP data that are not considered duplicates. All they have done is repackage the same directory information in different ways with different keywords and phrases.
There are thousands of sites using the same AP newswire articles that are not considered duplicates. All they have done is repackage the information.
I could go on.
I am considering doing a similar thing with my directory content and provide it to other sites. Are they all then to be considered duplicates? The only difference here would be me owning and operating the sites vs. different people owning and operating the sites. Either way, the majority of the content would be the same.
3cents, I get the feeling this is not as cut and dry an issue as you make it. Any other opinions? Maybe GoogleGuy can weigh in on this?
It's up to you how you want to do it. I am speaking from personal experience that duplicate content spread across multiple domains can get you in big trouble, and that building a site filled with original content really works well.
So far, the wisdom I've been able to discern is that this sort of content, being wrapped in different navigation and structure, is not at risk of breaching the "substantially duplicate" admonition.
Also, I seem to recall Brett saying that product descriptions provided by manufacturers that go out to all distributors, retailers, etc. and show up in duplicate on multitudes of websites do not run afoul of the substantially duplicate risk.
This area is fairly nebulous, and any analysis of content which covers the same territory for its susceptibility to being a duplicate, is bound to be subjective to a large extent.
I've seen conjectures that anything less than 80% or 50% similar is OK. Take your pick.
BTW, my3cents, welcome to the boards. One thing I might mention, and I hope you'll take this constructively, is that my eyes glaze over when I try to read your posts. It would be very helpful if you could break them into paragraphs, as I feel you have much good input to contribute here and I'd like to read it.
My whole point about this is not if google will ban you for it, people with duplicate content can probably slip under google radar, simply because of the legitimate reason for duplicate content that go60 pointed out.
My point is that, if you do this, you stand a risk that someone will report you and spreading content across multiple domains, duplicating content and tweaking it to target a different set of search term, makes you a target and puts you at risk. Google may never find this, but there is a good chance that your competitors will.
If you want to do it go ahead, but you are running a pretty big risk. I'll tell you right now that if I find you, I will report it. I'm not looking for you, but someone else may be. If you are duplicating your content with the intent of increasing PR or increasing SE position, you are breaking the rules, and once someone reports you....
Make whatever choice you want, if you want to justify it, go ahead. If you want to talk yourself into believing that you're doing nothing wrong, fine.
I am telling you from experience, that you are gambling. If you are comfortable gambling, that is your choice, but the actions you described in duplicating content are against the rule, and if someone reports you, you will probably get hurt pretty bad.
I wish that everyone who is looking far ways to cheat the system would realize that substantial original content will get you so much further. Seriously, it's not some kind of joke, it really works. Why run the risk? you could just develop a few dozen new pages a month on your main website, make it keyword rich, and target the content and title to the search terms you want to target, you really will do better than the duplicate sites. Why would anyone want to run the risk of a ban when there is a better way that has been proven a thousand times?
One of my clients has a French section for his web site. He has also bought a domain name (in French) so that he can refer people to it. On the French domain name, he only has the French home page. All the links then point to the French section on the English domain.
He does this because he can't afford having 2 secure keys for the shopping cart.
Thus the problem: The French home page is duplicated.
In September his English side got a PR 0 (white) from Google. I found the duplicate content and put up a robots.txt on the French domain so that Google would not spider it.
This was done December 12th 2002. Since then, he still has a PR 0 even though there are good sites linking to him.
My question is: Will the robots.txt keep Google from the French domain site, and will it be enough to get the penalty removed from the English site?
Sure there is.
I've had extensive experience with dupe content over the last year. Sites being moved, sites being pure mirrors - etc etc.
I know one guy with 75 domains that are probably classified as duplicates by Ink and Google. fooatNewYork.com, fooInMiami.com, fooInStLouis.com - etc etc. The only difference is the map to find their store on the home pages.
Google thinks they are dupes. So What? One just gets a lower pr than the next one. No penalty, no drops from listings, etc etc. I've never seen one removed from the index for dupe content.
Dupe content can be a friend if used appropriatly.