I know, I know, this comes across as an "eye for an eye" type of deal but what I'm actually looking for is more like:"what would
Je..., I mean,
tedster do ?" and has a purely practical application.
Having managed to have most of my money earning sites banned in one day (
long story [webmasterworld.com]), I am looking at some austerity measures the likes of which I never had to implement before. It includes elimination of hosting accounts and aggregation of sites on a small number of VPSes which is where the question of the tremendous burden that Googlebot places on any site comes in.
Sure, there are other bots that sap bandwidth but Gbot deserves an honorary mention. It has always been the largest consumer of bandwidth and CPU but now, with real people from Google gone, the amount of resources it consumes compared to the benefit to me (ZERO) is simply ridiculous.
Despite the ban, Googlebot continues to pummel all my banned sites as if nothing happened. In fact, the amount of
URLs in web index in WMT continues to grow even after the ban.
So, what would you guys do: let Googlebot crawl the sites unchecked despite them bringing no traffic or ban Googlebot (either through robots.txt or possibly even firewall) which will then enable you to save money on smaller server/bandwidth.
There is another facet to this: I am using AdSense on most of my newly banned sites. Since
Mediapartners-Google bot shares the data, Google will still have a pretty good idea about new URLs on my sites even though Googlebot itself is banned. If and when any of my sites get un-banned (does that even happen?) I can re-enable Googlebot access.
So, does anyone think it may be a good idea to mess with Googlebot access?
Cheers!