Forum Moderators: open

Message Too Old, No Replies

Googlebot http 1.0 vs 1.1

What does it imply?

         

Spine

7:21 pm on Dec 11, 2004 (gmt 0)

10+ Year Member



I have a few sites on a virtual server setup. Each site has it's own IP, is running on the same OS and uses plain vanilla HTML.

On the majority of my sites, I get visits from the Googlebot that requests files using http 1.0, but on a particular site I see Googlebot with http 1.1.

Is there a particular reason why this would be, and given the choice, which one would your rather have crawling your site?

I've also noticed, when using a certain 'whois' service, that with the detailled results they return, all my sites are giving the 'whois' bot a 206 response code, while some of my competitors sites are returning a 200.

Using a server header checking tool, my sites return a 200, and I see visitors (including Googlebot) get a 200 as well. I only wonder because some of my competitors sites return a 200 code using the same 'whois' tool, could it be because they are using an older version of Apache that doesn't deal with 'partial get' requests?

Thanks in advance for any thoughts on this.

ciml

10:24 am on Dec 13, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



The main confusion is to do with name based hosting.

Googlebot, even though sending HTTP/1.0 GET requests, does send a Host header so you should have no problem.

bull

12:29 pm on Dec 13, 2004 (gmt 0)

10+ Year Member



AFAIK only the Mozilla/5.0 (Googlebot...) does use HTTP/1.1 . For Differences see also [webmasterworld.com...]

Spine

10:04 pm on Dec 13, 2004 (gmt 0)

10+ Year Member



Thanks for replies.

I'm a bit confused what you mean about how this relates to name based hosting, but I'm looking into it.

I have unique IPs for all my sites, but it does say somewhere in the control panel my host type - 'name based'.

From what I'm reading on the Apache site, it's easier to have name based unless you need ip based for SSL etc.

bumpski

4:55 pm on Dec 14, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



If your site supports GZIP compression of web page content it's likely we'll all start to see the 1.1 Googlebot. This new bot is crawling pages and requesting GZIPed or compressed content. It could save 3 X bandwidth if Google used it and Web host's would generally support GZIP compression.

Look at your logfiles. You may see the byte count on your pages is much smaller than usual.

All common webservers (APACHE, IIS) support dynamic GZIP compression now, but many, many web hosts do not turn it on (webhosts; your customers and their customers want GZIP). 56K modem users would benefit tremendously as would Google, it could crawl the web 3 times faster. Web page text would load 3 times faster!

Web servers and therefore hosts also support precompressed web pages, which are a great potential way to "cloak".

Googlebot could be checking for cloaking.

For a lot more see this link:
[webmasterworld.com...]