Forum Moderators: open

Message Too Old, No Replies

Two similar sites for spelling differences. Good or Bad?

Is this technique going to get me penalised (penalized)?

         

Philiboy

2:27 pm on Jul 22, 2003 (gmt 0)

10+ Year Member



I've copied my site to a .co.uk domain and a .com domain but I use PHP to test which domain the visitor is on. If .com all spelling uses American Engish spelling (optimization, penalization, color etc.) otherwise UK English (optimisation, penalisation, colour etc.)i.e. I call a function when each word occurs, so the 2 sites are not actually identical.
I do this so my site is optimised (alright, optimized!) for a UK audience and an International English speaking audience. I know there have been a lot of posts on duplication, but I still feel I need to ask whether people think I will get away with this. I don't do it to spam but for the legitimate reason of expanding my market (and I'm optimising for a larger variety of keywords). Unfortunately my domain names (apart from the .com / .co.uk suffix) are the same so I guess the duplication is going to be easily detectable. Should I generate different titles for each similar page? Should I make sure there are no duplicate pages at all (at the moment there are)?

Should I perhaps move one of the two sites to a totally different domain?

If it is going to be a problem, can anyone suggest another approach which will have the same effect?

ciml

3:34 pm on Jul 22, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Your .com domain and your .co.uk domain are different domains, so I wouldn't worry about that.

With such small changes between the pages, I expect them to be listed individually (not merged as identical pages would be), but for only one to show for any given search unless you click "repeat the search with the omitted results included" or use &filter=0.

So IMO the big question is would the matching spelling be the URL that's listed, and would it rank higher than the other spelling would if it was the only one? I suspect that the answer to both questions is yes.

Philiboy

2:13 pm on Jul 23, 2003 (gmt 0)

10+ Year Member



Thanks, Ciml. So you don't think I'll get penalised for near duplicate content? Some of the pages on my site will be identical as they happen to not use words which are spelled differently in American English compared to UK English. Does anyone think I should force all pages to all be subtely different to avoid penalisation? Is it advisable to make meta tag areas all different? How far should I go? I want to get my pagerank up which is only 2 on each site. Has anyone tried to handle spelling variants like me and if so, can they help me learn from their experience?

peewhy

2:27 pm on Jul 23, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



You have two sites with very minor differences, Google will drop one and index the other.

I would suggest you revamp the .co.uk so it is totally different in design and body copy so Google treat them both as different sites.

Philiboy

3:20 pm on Jul 23, 2003 (gmt 0)

10+ Year Member



Thanks, peewhy. Maintaining 2 sites is a huge overhead, though I guess I might need to be realistic and consider this. I was just wondering whether I could just go a whole load further with my PHP and just generate more differences

e.g. (ignore syntax) if (international()) { this sentence says one thing in one way } else { a different way of saying the same thing }

but not having to write everything twice in different ways, doubling my work.

I could do this in the meta title area, keyword area etc. not just to make the spelling differences but also to make the sites different enough to avoid being penalised. But how far do I need to go to avoid penalisation?

The 2nd purpose of the PHP is that the .uk can include info only relevant to the UK and it is omitted from the .com site. I maintain one local copy of my site and don't have to maintain two variants which requires a lot of management. The differences are handled dynamically via the php.

Remember, no deception or attempt to spam is intended in this approach - it's so people looking for (for example) phrases such as "search engine optimisation services" will find me as well as if they type "search engine optimization services" (note different spelling).

Herenvardo

3:50 pm on Jul 23, 2003 (gmt 0)

10+ Year Member



I can suggest two solutions, but both of them might have some problems:

1st: Why don't create a only site with dinamicall content? You should be carefull to keep your PR and you on-page factors for SERPs, but it might work

2nd: Use standard English. Standard English is that variant of English that nobody uses. It exists only to provide you with a standard to solve cases like this. In countries where, at school, English is learned as a foreign language, that variant is the one that is studied. It is a problem for students like me, we learn an English that doesn't really exist, but I think that this is the case where standard could be useful.

Philiboy

4:28 pm on Jul 23, 2003 (gmt 0)

10+ Year Member



Thanks Herenvardo. It is effectively one site (but copied to two places) with dynamically created content (via PHP programming). It is not database driven, but it is dynamic.
The PHP detects whether its xxxx.co.uk or xxxxx.com.
Not sure about what you are saying about standard English. I need to reach my UK audience (one spelling e.g. colour) and my International audience (U.S. and anyone else using American spelling e.g. color). That's actually the point.

peewhy

5:46 pm on Jul 23, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



There is a very simple solution, a good generic welcome page with two flags for US & UK each linking to their relevant pages and only one site :)

ciml

7:55 pm on Jul 23, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Philiboy:
> So you don't think I'll get penalised for near duplicate content?

I wasn't considering human penalties in that post, just what Google's software would do with those pages.

peewhy:
> You have two sites with very minor differences, Google will drop one and index the other.

An interesting notion. Would this be due to human intervention, or because the URLs would be merged as identical?

peewhy

7:59 pm on Jul 23, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Google will see one of them as a mirror and not index it.

Herenvardo

4:53 pm on Jul 29, 2003 (gmt 0)

10+ Year Member



About similar URLs, I've seen duplicated sites indexed twice in google search results. It happened with a site called www.domain.es. After it appeared another one, exactly identicall but called www.domain.net. Since this site was spamming and it was compitance, i reported it to google. So I can guess that duplicated pages will only be punished if somebody reports them as spam.
Even I can't ensure it, I hope that your site won't be punnished. If you let your visitors know that they are in the UK site and there is an American version (and vice versa), you will dificulty get reported as spammer and, even so, when checking the page, google guys will see that there is no duplication.

Greetings,
Herenvardö

vincevincevince

5:30 pm on Jul 29, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



i say point them both at exactly the same site (301 the .co.uk?)

THEN sniff for country by IP block - and return the spelling that suits.

using .com and .co.uk to determine someone's country is certainly not a good solution!

ITcameleon

5:59 pm on Jul 29, 2003 (gmt 0)

10+ Year Member



Certainly DO NOT place pages in British English and American English on one site. I had a welcome page with two flags and... well my site got penalised and I had to merge the pages.
I'm also considering creating a UK version and disallowing googlebot using robots.txt The site in question is in html
It would be better to have different versions for UK/USA - different colour schemes and graphics, different price lists(!) etc.
So how do you think - should I simply create two versions and disallow the spiders from one?
Regards.

alpine

7:07 pm on Jul 29, 2003 (gmt 0)

10+ Year Member



I've got the same issue right now.

We've created localized versions of our site, in local language, local currency prices, and with a ccTLD domain.

So for a country like Italy, or any with a non-english language, I'm not worried about a dupe content penalty, because the language displayed will be different than our .com site.

But for .co.uk, the only difference will be that prices will be in pounds.

I want to be sure our .com will not disappear from the SERPS in favor of .co.uk!

glengara

7:24 pm on Jul 29, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



To a certain extent, this is where PR comes into the equation, the more you have the better one page can do for both spellings.
If G shows a "did you mean..." for your KW, even one instance of the Brit spelling in an otherwise US spelling may well be enough to rank you for both.

RobinC

7:27 pm on Jul 29, 2003 (gmt 0)

10+ Year Member



I'd say point them both to the same place too, in my case I point everything at my hostname, and only mention my .com (in my .sig etc)... If you wanted to do subtle differences for spelling then I'd suggest using the browsers supplied accept language header (_SERVER["HTTP_ACCEPT_LANGUAGE"]), parse that, and see which comes first - "en" or "en-us". You should be able to do a small php function along the lines of <?php english("british spelling","american spelling");?> - that way google will only have a single site to index... Of course, that then makes me wonder what language (if any) do the spiders ask for ;-)

On the other hand, if you had the time, and there was enough business involved in a seperate .uk page, then doing different content should keep everyone happy, not least of which your bank manager...

Vec_One

9:36 pm on Jul 29, 2003 (gmt 0)

10+ Year Member



My situation is somewhat similar to Alpine's. We have many different language versions on the same domain. Although most pages are unique, some pages (prices, schedules) are very similar.

Can anyone give me an estimate as to how much content should be different?

Philiboy

10:56 am on Jul 30, 2003 (gmt 0)

10+ Year Member



Thanks for all the posts. Here are some comments :

vincevincevince (..sniff for country by IP block..) : I never thought of that. Excellent idea. I'd have to figure out how to program it. Know of any example? By the way, I'm not really figuring out which country via .co.uk versus .com, because I make it clear on the site that the .co.uk is dedicated to the UK version and there's a link to click to the international version. Note there is no U.S. version - it's international (outside the UK) and UK.

ITcameleon (..also considering creating a UK version..) : If you want to go down this route (i.e. I'm taking all this advice on board and may reconsider), try using php so each version is created dynamically. I'm keeping my .php secret, but it is not very difficult. Hint : strstr($_SERVER["DOCUMENT_ROOT"]. Write a few simple functions based on that value. Just wrap PHP around the bits that need to be different. Force each page to be different and make sure each pair of meta titles, meta descriptions, meta keyword areas and main body are different. Much better than having to maintain two versions of every file.

alpine (..I've got the same issue right now...) : I also use my .php technique to keep the info that is only of use to my UK audience limited to the .co.uk site.

glengara (..this is where PR comes into the equation...) : Very good point. Don't think it will work 100% because if someone's search involves, say, "optimisation" spelled the UK way and they get heaps of results (which they often do) they probably won't click on the "did you mean" results, especially if they're mainly interested in UK sites anyway.

RobinC (..I'd suggest using the browsers supplied accept language header..) : But I want people to find me on either spellings when they do a search. Being listed for only one spelling variant won't help me. Useful info though!

Vec_One (..how much content should be different?..) : See my reply to ITcameleon. Search engines could save a checksum for each area of your page (checksums are unique per content) and if they spot two matching checksums they'll flag duplicate content, investigate and may penalise you. I'm not saying that they do this (only they know), but I'm guessing that they could do now or in the future.

Thanks for all the advice. I'm pursuing my .PHP technique for now (but I've forced more differences) and it remains to be seen how my pagerank will be affected. Probably a Google Dance this weekend. At some point I might switch to vincevincevince's suggestion if someone could advise me on how to program this?