Welcome to WebmasterWorld Guest from

Message Too Old, No Replies

Using Canonical tag on search results refine pages

4:06 pm on Jan 29, 2011 (gmt 0)

Junior Member

10+ Year Member

joined:Jan 10, 2005
votes: 0


We recently had a big drop of traffic to our search pages (December 29) and have yet to recover. We are considering using the canonical tag to help solve what we believe is an internal duplication problem.

Here is the scenario:

1) We moved to a much bigger server in November, which resulted in Google indexing tens of thousands of "refine" style search pages.

So, for example, we have a primary search page which previously ranked well:


However, we also have links which basically refine the search by category, such as:


Although the 2 pages are not "identically" duplicate (they pull out refined results on the same keyword but have different on-page text and meta data), would it be wise to use the canonical tag to point Googlebot to the main search page (Google has indexed tens of thousands of these similar pages and we are wondering whether this has contributed to the drop in rankings)

2) Just to add more complications, we have a 2nd search link:


which links from /search1/ page and the structure of this page works in a similar way regarding refine URLs off it.

On our previous server, we think because of the relatively low crawl bandwidth available to Googlebot, they weren't indexing so many of these refine type pages, and therefore we avoided internal duplication?

Will the canonical help in these cases? And do you think this has been the major cause in dropped rankings?

Any help is appreciated.

6:58 pm on Jan 29, 2011 (gmt 0)

Senior Member

WebmasterWorld Senior Member tedster is a WebmasterWorld Top Contributor of All Time 10+ Year Member

joined:May 26, 2000
votes: 0

This seems like a type of faceted search result. The canonical link seems like a pretty big stretch to me in these situations, but I know sites are doing it.

I prefer not to allow faceted searches into the index at all - just one version. The crawling has got to be more effective as I see it.
7:57 pm on Jan 30, 2011 (gmt 0)

Junior Member

5+ Year Member

joined:July 24, 2010
votes: 0

Aren't search results supposed to be noindex?
8:37 pm on Jan 30, 2011 (gmt 0)

Senior Member

WebmasterWorld Senior Member tedster is a WebmasterWorld Top Contributor of All Time 10+ Year Member

joined:May 26, 2000
votes: 0

Yes, if they are true site search results. But many sites are built on a PHP/MySQL framework or something parallel, so any category page is, strictly speaking, a database search result. But that is part of my thinking is recommending noindex for any faceted search or re-ranking your site can generate.

I would, however, suggest not using the word "search" in the URL itself. And even more, work to get some kind of direct navigation in place (drill down browsing, rather than search).
1:27 am on Jan 31, 2011 (gmt 0)

Moderator This Forum from US 

WebmasterWorld Administrator robert_charlton is a WebmasterWorld Top Contributor of All Time 10+ Year Member Top Contributors Of The Month

joined:Nov 11, 2000
votes: 165

The choice of categories also needs to be limited and strategic. It's not just crawl bandwidth you need to be concerned about... it's also PageRank distribution and a clear navigation hierarchy.

If you have a hotel site, you probably want to categorize for Google by city. You probably don't want Google to index categorization, say, by smoking vs no smoking or by price.

The navigation structure needs to reflect those priorities so that the pages which you choose for Google to index are also structurally your most important pages.

Even with the robots="index,follow" meta tag, you should avoid building a structure that puts your noindex pages up near the top of your hierarchy. And do not use rel="nofollow" to shape PageRank flow.

Join The Conversation

Moderators and Top Contributors

Hot Threads This Week

Featured Threads

Free SEO Tools

Hire Expert Members