Someone refresh my memory, please: Why are we supposed to not block BlueCoat ranges?
199.19.249.196 - - [15/Sep/2015:01:11:25 -0700] "GET /fonts HTTP/1.1" 301 593 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0"
199.19.249.196 - - [15/Sep/2015:01:11:26 -0700] "GET /fonts/ HTTP/1.1" 200 8261 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0"
199.19.249.196 - - [15/Sep/2015:01:15:02 -0700] "GET /hovercraft/april_blues.html HTTP/1.1" 200 277436 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0"
199.19.249.196 - - [15/Sep/2015:01:16:12 -0700] "GET /hovercraft/nunavut99 HTTP/1.1" 301 623 "-" "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0"
et cetera for a total of 418 requests. I don't think this site actually has 418 pages; the total was bloated by all those slashless requests.
Also a request for /paintings// with spurious extra slash, which makes me wildly uneasy because the Googlebot has also been requesting this lately, and I swear I can't find any malformed links, or else the other search engines would be requesting them too. Mostly pages, except a few non-page files that happen to have <a href> links.
User-Agents toggled between
Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:40.0) Gecko/20100101 Firefox/40.0
and
Mozilla/5.0 (Windows NT 6.1; WOW64; rv:40.0) Gecko/20100101 Firefox/40.1
-- mostly the latter.
No robots.txt, but plenty of requests for pages in one roboted-out directory (one of two, and the less visible one at that). Further weird quirk: they didn't get around to requesting the front page until about a quarter-hour into the visit. Final weird quirk: although they didn't bother about robots.txt, they
did ask for the sitemap. I consider that rude.
Checking back in my records, I find that this particular range-- 199.19.248.0/21 --was blocked at one time, later unblocked due to apparent humans ... and is now decidedly blocked, in case someone comes rattling the barn door in search of additional horses.
....
Oh, ###. As I write this, I realize that an additional, seemingly unrelated robot from the next day-- which I've been avoiding taking a closer look at because it's too ### complicated, having done a fine job of impersonating a human-- was using the identical two User-Agents as this unwanted BlueCoat visitor. WTF?