Forum Moderators: Robert Charlton & goodroi
Thanks
BUT! There is a small "but" always :) I guess you are making too harsh decision. Surfers like WWW. Tomorrow someone will post an advice somewhere.. he is a loyal customer of you... he remembers you URL... but he knows that site's name always goes with www. You lost those natural links. Look at the boss. E.g. [google.com...] . Why do you not follow that way?
[edited by: tedster at 3:45 pm (utc) on Jan. 21, 2006]
[edit reason] use example.com [/edit]
URL Console has unexpected behaviour: IF YOU REMOVE ONLY NON-WWW URL OR ONLY WWW URL IT REMOVES BOTH! So trying to solve www/non-www duplicate problem with URL Console leads directly to six months of your site out of the index!
[edited by: tedster at 4:28 am (utc) on Jan. 21, 2006]
[edit reason] change formatting [/edit]
1. Note: If you believe your request is urgent and cannot wait until the next time Google crawls your site, use our automatic URL removal system. We'll accept your removal request only if the page returns a true 404 error via the http headers. Please ensure that you return a true 404 error even if you choose to display a more user-friendly body of the HTML page for your visitors. It won't help to return a page that says "File Not Found" if the http headers still return a status code of 200, or normal.
AND
2. Note: If you believe your request is urgent and cannot wait until the next time Google crawls your site, use our automatic URL removal system. In order for this automated process to work, the webmaster must first insert the appropriate meta tags into the page's HTML code. Doing this and submitting via the automatic URL removal system will cause a temporary, 180-day removal of these pages from the Google index, regardless of whether you remove the robots.txt file or meta tags after processing your request.
As you can see there is a difference betweeen 404 and robots restrictions.
Wizard, keep your horses. You return 404 on www requests and 200 on no-www requests - it works. If you are in rush, e.g. some sencitive information got in SERPs by mistake, you need to use console. By the way, there is nothing about temporary removing if you use 404.
In the past, I used meta robots 'noindex' for one URL version while 'index' in another, and it removed both. I also removed an outdated domain with robots.txt and so I know it also removes both. I didn't check what if one version returns 404 and another 200 - if this works as expected, would be ok.
Tedster, sorry for the formatting, I shouldn't have bolded all my post, you're right :) I reacted too emotionally seeing that someone is likely to unintentionally harm their site ranking just because of this know URL Console issue.
seeing that someone is likely to unintentionally harm their site ranking
Understood. And I agree with the urgency.
Most definitely, do NOT try to fix canonical issues by using the URL Console to remove just one version (whether with or without the 'www'). Matt Cutts also mentioned this quite strongly. The console will just zap everything for you.