Forum Moderators: Robert Charlton & goodroi
I am using robots.txt for blocking affiliate pages of my site.
do you think that Blocking more and more pages in robots.txt for Googlebot, is it creating a problem for my site?
Due to this problem, is it possible that my ranking in Google will be gone down?
Because there is affiliate urls are increasing more and more and for protecting my site from the penalty of Duplicate content i have blocked all affiliate urls in robots.txt and even any new affiliate url generate that will also block in robots.txt
I am using this command for blocking my affiliate urls:
User-agent: *
Disallow: /*?a_aid
Disallow: /*?
Because before implementing this command in robots.txt my ranking is very well but after blocking these kind of pages in robots.txt, I found that my ranking gone down.
so now what i should to do for getting my ranking up.
Let me know your suggestions.
Thanks
Blocking urls in robots.txt does not cause problems with Google just because you do it. But if your rules do something other than what you intended, then there can be big trouble. If that last Disallow rule I pointed out is really what you need, then the drop in your rankings may have a different cause.
At the same time, if Google previouly passed link juice through all those backlinks and you sort of chopped them off from receiving and circulating PR - that could create some drops. Another approach is to take any affiliate ID, store the information you need to do right by your affiliate, and then serve a 301 redirect that removes the query string. This way you allow Google to give you the credit for those backlinks and to pass along any link juice.
Robots.txt working well and it blocks only those urls which contains "?".
and as you suggest that if i will give 301 redirect to all urls which contains "?" then i will get the benefits of Link Juice.
But i am getting confused. if i am giving 301 redirect on affiliate urls, then affiliate tracking is possible? or it creates problem in affiliate order tracking?
Let me know...
Yahoo and MSN also support wildcard pattern matching in robots.txt.
Sure, and so they should also get added to the list of explicit bots in that robots.txt - if thats all the site owner cares about then it is better to disallow all other bots than suffer unavoidable anger that some other bot that is perfectly robots.txt compliant won't obey these non-standard directives.
Hell, no. I do it for a variety of reasons for many pages and it works just fine. I mean do I really want Google to index the page that serve the alternate ads from? It has no content except code for Amazon ads.
<<Due to this problem, is it possible that my ranking in Google will be gone down?>
No, absolutely not. Actually Google only wants to index pages that make sense to humans. The page that I have blocked with just Amazon code is really not meant for a human visitor directly.