Forum Moderators: Robert Charlton & goodroi
One item which is missing from this list is disallowing crawling of duplicate content with your robots.txt file. We now recommend not blocking access to duplicate content on your website, whether with a robots.txt file or other methods. Instead, use the rel="canonical" link element, the URL parameter handling tool, or 301 redirects. If access to duplicate content is entirely blocked, search engines effectively have to treat those URLs as separate, unique pages since they cannot know that they're actually just different URLs for the same content.
I've read that Google might give less trust to a site that has an excessive number of redirects.
But whatever you do, it might be best to do it gradually if you can.
I've read that Google might give less trust to a site that has an excessive number of redirects.
In some cases it's an indication of major revamping, which could be a long-term negative mark against a site.
There *could* be some PR juice loss though (as there also is by using 301).
[edited by: Sgt_Kickaxe at 7:39 pm (utc) on Apr 25, 2011]
- Robots.txt will tell Google which pages cannot be indexed.
Another option is to have a single canonical url for the product. Allow the users to change the color and such through Javascript and AJAX within the page.
What to do if you have 3 versions of same URL, served through same php file.
example.com/green-Blue-Widgets.html <--- look at small letter for word Green. I don't know how google pulled that out.Are you on IIS server? They are case insensitive.
example.com/Green-Blue Widgets.html <--- See the space between Blue and Widgets