Forum Moderators: DixonJones
>> method of tracking googlebot's crawling of a Web site
There are a variety of log analysis tools available to review logs. I use NetTracker because it does allow me to follow a particular user agent (a bot registers a user agent just like a browser does).
>> Is Googlebot a focused crawler?
Not sure what you mean by this. Googlebot can be found under a variety of IPs and there are two versions of the bot: Fresh and DeepCrawler. A search here on WebmasterWorld should turn up more than you ever wanted to know about Google - and read this [searchengineworld.com]. ;)
A focused crawl is where the robot enters a site looking for pages on a particular theme. Lets say the inbound link's anchor text says "log analyzers." Focused crawlers would only follow links that are relevant to those keywords, although each search engine may have different "focuses" for a visit. I wonder if Freshbot is a focused crawler and what its mission is each time out.
According to some postings I remember seeing last year, some Webmasters seem aware of which pages the crawlers were picking up. So I'm wondering if any of them have some insight into the robot's purpose for a specific visit and which pages get picked up.