Forum Moderators: open
And Now Google's Doing It. JS Stats Show GoogleBot
Given one page to examine isn't really a crawl...
[edited by: TheMadScientist at 7:24 pm (utc) on May 17, 2011]
They may be claiming that's what they're not doing, but lucy24 noted in this thread previews were activated and 100 requests were made, including requests for linked hypertext documents and the files included on those pages, which makes it a 'crawler', and subject to 'protocol' imo.
Prefetch is still not a crawler, it's a browser function
[edited by: TheMadScientist at 7:57 pm (utc) on May 17, 2011]
I was thinking we were still on the 'Web Preview Bot', which I don't think could be considered a browser function, but you said 'prefetch'.
Because web preview bot would require a browser, and I didn't make the decision, blame FireFox and Google for nasty old prefetch.
[edited by: TheMadScientist at 7:44 pm (utc) on May 17, 2011]
For instance, my entire site is NOARCHIVE, wouldn't a screen shot qualify as an archive?
As on-the-fly rendering is only done based on a user request (when a user activates previews), it’s possible that it will include embedded content which may be blocked from Googlebot using a robots.txt file.
No. You must show Googlebot and the Google Web Preview the same content that users from that region would see (see our Help Center article on cloaking).
You can block previews using the "nosnippet" robots meta tag or x-robots-tag HTTP header. Keep in mind that blocking previews also blocks normal snippets. There is currently no way to block preview images while allowing normal snippets.
No. You must show Googlebot and the Google Web Preview the same content that users from that region would see (see our Help Center article on cloaking).
[edited by: TheMadScientist at 11:46 pm (utc) on May 17, 2011]
I could cause 100 web preview hits easily enough: set number of results/page to 100 and do a site: or whatever search.
it is not technically a bot. It's a proxy.
I could cause 100 web preview hits easily enough.