Forum Moderators: goodroi
robots.txt is handy for blocking certain directories or search engines that you don't want or care to be in.
If a bot uses 100mb of bandwidth each month to crawl and only sends you 2 referrals, you might want to block it.
It is also a good tool to tell the bots not to index certain sections of the site.
If you have a members only area you don't want that in the SERP's
If you use dynamic content (a page has more than one URL) you could block all but the proper URL for that page.
Block click tracking scripts.
Block pages that are nuthing but java.
maybe there is a page that you just don't want in the SERP's for some reason.
If you have pages that are so similar (red-widgets and reddish-widgets) that you are afraid of duplicate content penalty but (for the user) you really want both pages, you could simply block one to avoid a possible penalty.
So with use of robots.txt on a good robot (in major SE's) IMHO I believe that - yes- robots.txt can help improve ranking.
Bad bots - there are lots of bad bots (or undesirable in your niche) that do obey robots.txt
for these there are so many just deal with them as they come, each new bot that shows up just do a little search to see what it is and decide to block or not. Some bots you may want to block from certain directories while others may be a different case.
Basically:
//Instructions for all robots who care
User-agent: * User-agent: msnbot Disallow: /scripts/ Disallow: /*.js$ Crawl-delay: 120