homepage Welcome to WebmasterWorld Guest from 54.227.77.237
register, free tools, login, search, pro membership, help, library, announcements, recent posts, open posts,
Become a Pro Member

Home / Forums Index / Google / Google News Archive
Forum Library, Charter, Moderator: open

Google News Archive Forum

    
How to permanently Kill your duplicate domain
Thoughts?
johnser




msg:130186
 1:13 pm on Sep 18, 2003 (gmt 0)

2 PR6 domains (x.com & y.com) point at same IP.

There were problems in the summer with a 301 redirect.
Y.com is older and is being crawled regularly as it has been for the last 3 years.

We moved y.com to a different IP 6 weeks ago, removed all content & placed 1 link on every page saying "please click here" (which goes to x.com)

Y.com is still being crawled daily & x.com is not.

x.com is on the old IP. Y.com is a new IP.
Doing a G "site search" for X.com displays all y.com URLs.

Both sites are obviously regarded as the same thing by G & (both) have 100 backlinks

Whats the most bullet-proof method of encouraging Googlebot to not crawl y.com but visit x.com instead?

Thanks in advance
J

 

Mike12345




msg:130187
 11:55 am on Sep 19, 2003 (gmt 0)

How about placing a simple link from y to x. Accompanied by a message explainng that the site has been moved/deleted/whatever, so whenever a user sees the site they follow the link. Then block all bots by using robots.txt

Then email all folks that link to site y and them to change the link to site x, that way google will eventually stop being referred to x, and if it does get there it wont be able to index it.

I hope that helps :) its a really simple way but i find when keeping things simple it often works better.

:)

coco




msg:130188
 1:13 pm on Sep 19, 2003 (gmt 0)

301 redirect Y to X

johnser




msg:130189
 4:12 pm on Sep 19, 2003 (gmt 0)

Thanks for that.

Mike - A simple link is in place from every page.
All links point to X.com yet Y.com was being indexed?
G was ignoring our robots.txt.

Coco - We tried that during the summer but reading on WW, it seems G was having bad problems re permanent 301 redirects.

J

xlcus




msg:130190
 12:06 pm on Sep 21, 2003 (gmt 0)

I've used a 301 redirect to move a site a couple of times, one in the last month, and once about 3 months ago. Both times, the old site was dropped and the PR transfered without a problem. Perhaps you could give it another try?

dirkz




msg:130191
 12:11 pm on Sep 21, 2003 (gmt 0)

If you use a 301, do you use a RedirectMatch with a regexp for all requests (RedirectMatch permanent /.* ...)? Or do you just redirect "/" and thus let all other pages issue 404s?

Just wondering whether it could be a problem for Google.

johnser




msg:130192
 11:03 am on Sep 22, 2003 (gmt 0)

I'm not sure of that dirkz.
I'll ask the techie & see what I find out.

Thanks all for your help.
John

bkuemmer




msg:130193
 2:57 pm on Sep 23, 2003 (gmt 0)

We have a very similar problem with a slightly different setup: We had two different virtual domains serving the same pages and we "split" them 6 weeks ago (i.e. some content stayed on y.com and some moved to x.com, with 301 redirects from y.com to x.com.

Especially
Doing a G "site search" for X.com displays all y.com URLs.
Both sites are obviously regarded as the same thing by G
applies in our case as well

Unfortunately, we have not yet figured out why this is happening and what to do about it. G has just now started visiting *one* page on x.com (which has very good inbound links), but apart from that only the server root is visited by googlebot.

jdMorgan




msg:130194
 3:45 pm on Sep 23, 2003 (gmt 0)

A 301-Moved Permanently [w3.org] response is the way to go. RFC 2616 [w3.org] describes this server response and what it means. If you have problems with it, then either it's not implemented correctly (check it [webmasterworld.com]), or you haven't given the search engines enough time to pick it up. There have been rare cases when a search engine does not correctly handle robots.txt, but that does not mean you should not do it correctly, and let them fix their problem (You could do a temporary work-around if you were able to infer the specific problem and come up with a work-around for it).

Similarly, if Google is not interpreting your robots.txt file correctly, then it's likely your robots.txt syntax [robotstxt.org] is incorrect, so check it [searchengineworld.com]. I've seen minor problems with other robots misinterpreting robots.txt, but Google's parser is one of the most sophisticated ones.

Do things "by the book" for best overall results and minimized headaches. Then apply work-arounds when and if necessary. Check your work and be patient -- it can take 60 days before some search engines catch on, and even longer for others. :o

Jim

dirkz




msg:130195
 3:45 pm on Sep 23, 2003 (gmt 0)

only the server root is visited by googlebot

This looks like the behaviour Googlebot shows with fresh sites, which just goes for some time once the site is deepcrawled. So to me all looks good, though it will take more time.

Global Options:
 top home search open messages active posts  
 

Home / Forums Index / Google / Google News Archive
rss feed

All trademarks and copyrights held by respective owners. Member comments are owned by the poster.
Home ¦ Free Tools ¦ Terms of Service ¦ Privacy Policy ¦ Report Problem ¦ About ¦ Library ¦ Newsletter
WebmasterWorld is a Developer Shed Community owned by Jim Boykin.
© Webmaster World 1996-2014 all rights reserved