homepage Welcome to WebmasterWorld Guest from 23.22.173.58
register, free tools, login, search, pro membership, help, library, announcements, recent posts, open posts,
Become a Pro Member

Home / Forums Index / Google / Google SEO News and Discussion
Forum Library, Charter, Moderators: Robert Charlton & aakk9999 & brotherhood of lan & goodroi

Google SEO News and Discussion Forum

    
Removing duplicate content as fast as possible
301 and sitemap combination?
itloc

10+ Year Member



 
Msg#: 31180 posted 12:56 pm on Sep 12, 2005 (gmt 0)

Hi there

I do have a real problem with my largest website. It's basically a large multilingual site containing thousands of widgets.

Now my problem is: multilinguality. Imagine there are two languages. In one language there are 10-thousands of widgets available. The other offers only a few thousands.

Now when a user in language 2 searches for a certain widget he gets only 5 to 10 results. I show a link where he can include widgets in language 1 to increase the amount of results.

Google read the results of these links in his latest deepcrawl - and voila ... lots of pages got dropped.

Example:

www.widgets.cm/lang1/widgets/widget3000.htm
(Original URL in Language 1)

www.widgets.cm/lang2/widgets/widget3000.htm (Lang 2)
(URL if displayed in Language 2 section)

The listing for Widget no. 3000 is available both in language1 and language2. All that changes is the language of the navigation and the additional functionality that is shown. The description of widget 3000 is still in language 1.

This setup is good for users which prefer language 2 because it makes it much easier for them to navigate.

Good for the users - but bad for the search engines.

Now, what happened:

Google read initially Widget 3000 in language 1. After a deepcrawl Google read Widget 3000 in language 2, found it to be dupe content and dropped both.

What I try to do:

Now each time somebody ( or Gbot ) is accessing Widget 3000 in language 2 i do a 301 to Widget 3000 in language 1.

Example:

www.widgets.cm/lang2/widgets/widget3000.htm
-->301-->
www.widgets.cm/lang1/widgets/widget3000.htm

Additional thoughts and my main question:

I do use Google sitemaps. For both language sections I do only list the widgets which are available in that language.

Now that I know that Google has read 10 thousands of widgets in the "wrong" language sections my target must be to deliver as much 301s as possible to fix that situation.

The question is: should I include my language2 urls which only point (301) to the according widget in language 1 in my sitemap files?

Any feedback is highly appreciated!

Thanks a lot and sorry for my rusty english - hope you understood what i meant.

itloc

 

nsqlg

10+ Year Member



 
Msg#: 31180 posted 4:40 am on Sep 13, 2005 (gmt 0)

Maybe the google removal tool can be fastest way. (Googleguy tell some words about, please, because dupe filter is a trap for webmasters that dont know what is SEO, so dont wanna fool G, please!).

Also, I will like if somebody tell your experience with google removal tool for solve duplicate content problem (and how quickly can be).

Thanks.

itloc

10+ Year Member



 
Msg#: 31180 posted 12:03 pm on Sep 13, 2005 (gmt 0)

Hi and thanks

I tried the removal tool for other urls - and it worked well. However, as it has been stated before - removed sites will come back.

Using the removal tool is not possible in my current case - just because the robots.txt file would become too huge. I have around 130'000 urls... and i can't use wildcards.

It would be very nice to get GG's statement regarding my issue. But I assume that it is not GoogleGuys job to give advice for individual site problems. If he would start to do so - he would spend most of his time here.

May be I can just suggest the following: The Google Webmaster Guidelines could be a little bit more specific - or Google could create an extended version for professional webmasters.

Regards

itloc

nsqlg

10+ Year Member



 
Msg#: 31180 posted 6:14 pm on Sep 13, 2005 (gmt 0)

I tried the removal tool for other urls - and it worked well. However, as it has been stated before - removed sites will come back.

How many weeks delayed to come back? The pages penalized by dupe filter back again normal?

The Google Webmaster Guidelines could be a little bit more specific - or Google could create an extended version for professional webmasters.

I agree.

Force the googlebot with sitemap to crawl these 301 pages will work I guess, but seems a bit slow.

Wizard

5+ Year Member



 
Msg#: 31180 posted 6:31 pm on Sep 13, 2005 (gmt 0)

You should translate widget description to language2, but I understand it costs money.

The alternative solution would be to put the duplicate description on lang2 page hidden in JavaScript document.write, but this leaves no unique content on this page anyway.

If you cannot manage to make unique content on a page, don't feed it to Googlebot at all - use noIndex or robots.txt to block such pages.

Global Options:
 top home search open messages active posts  
 

Home / Forums Index / Google / Google SEO News and Discussion
rss feed

All trademarks and copyrights held by respective owners. Member comments are owned by the poster.
Home ¦ Free Tools ¦ Terms of Service ¦ Privacy Policy ¦ Report Problem ¦ About ¦ Library ¦ Newsletter
WebmasterWorld is a Developer Shed Community owned by Jim Boykin.
© Webmaster World 1996-2014 all rights reserved