homepage Welcome to WebmasterWorld Guest from
register, free tools, login, search, pro membership, help, library, announcements, recent posts, open posts,
Become a Pro Member
Home / Forums Index / Google / Google SEO News and Discussion
Forum Library, Charter, Moderators: Robert Charlton & aakk9999 & brotherhood of lan & goodroi

Google SEO News and Discussion Forum

Duplicate URLs and time to return to previous rankings after removal
Our site was spidered with a query parameter and then went supplemental

 10:56 pm on Jul 26, 2007 (gmt 0)

Our site was recently spidered with the query strings appended to the regular URLs.

Now, the URLs with the query string are ranking about one page less than the previous URLs and the previous URLs are supplemental.

We took the step to block the query parameter URLs with our robots.txt in hopes that our main URLs return.

How long do you think this will take?

Is there anything else we should do?



 7:39 am on Jul 27, 2007 (gmt 0)

It should straighten out in one or two spidering cycles, however long that is for your site. I would not do anything more for two weeks or so, at least.


 5:55 pm on Jul 27, 2007 (gmt 0)

Ted, thanks for confirming. It's a little surprising how sensitive google was to this change.


 9:14 pm on Aug 2, 2007 (gmt 0)

If this were to happen in the future, are we better off 301ing to the correct URL, or using the robots.txt to prevent it from getting crawled?


 12:29 am on Aug 3, 2007 (gmt 0)

That's very hard to answer without a knowledge of how query strings work on your site. As a general rule, stop the server from resolving dupe urls. Then a 404 cures many ills. If there are important backlinks to the dupe url, then 301 if you can't get them changed. If you don't need any query string urls in the search engines, block them all with robots.txt. Depending on your schema for generating query strings, you may be able to block just some of them with robots.txt and allow others.

Whatever you do, create a work flow discipline that prevents future issues.


 3:08 am on Aug 3, 2007 (gmt 0)

I once had someone do me the "favor" of linking to a site of mine with spurious query strings, intending to cause just this kind of dupe problem. So I temporarily lost the ranking of a few already-low-ranked pages, but in response, I quickly 301-redirected all of those bogus URLs to the correct URLs. Thanks for the PageRank!

It didn't last long of course, but keeping the redirect in place prevented any further such silliness.



 6:02 am on Aug 3, 2007 (gmt 0)

This whole querystring causing duplicate content issues is a MAJOR issue. What if you were running ad campaigns with legit sources which pass you a querystring to show affiliate information, or campaign information?

Or better yet, what if you have a way for customers to sort information on the page and you're passing a string to define the sort option.

Google really has to do something about this.

Robert Charlton

 6:22 am on Aug 3, 2007 (gmt 0)

Google really has to do something about this.

I'm not disagreeing, but I think Google is generally handling inbound query strings much better than Yahoo and MSN are. I can routinely expect problems on both of them, whereas problems I've seen on Google have been rare.


 7:45 am on Aug 3, 2007 (gmt 0)

Jd has a good point. It's almost always better to 301 redirect as opposed to using robots.txt in this case to 'guide' Google as to what credit belongs to which page.

Global Options:
 top home search open messages active posts  

Home / Forums Index / Google / Google SEO News and Discussion
rss feed

All trademarks and copyrights held by respective owners. Member comments are owned by the poster.
Home ¦ Free Tools ¦ Terms of Service ¦ Privacy Policy ¦ Report Problem ¦ About ¦ Library ¦ Newsletter
WebmasterWorld is a Developer Shed Community owned by Jim Boykin.
© Webmaster World 1996-2014 all rights reserved