Forum Moderators: DixonJones
It did download robots.txt before going ahead, but it has no URL or other contact info, and I can't seem to find more about it at cluster.ischool.washington.edu which is who the IPs belong to.
I decided to deny access to it, until I learn more about it. This is from the logs (just one site, it's accessing several sites on different IPs, so it must have ample bandwidth, or really-really like me ;-)
128.X08.131.220 - - [07/May/2004:18:54:45 +0300] "GET /****.html HTTP/1.1" 403 235 "-" "WIRE/0.1 (Linux; i686; Bot,Robot,Spider,Crawler)"
128.X08.131.220 - - [07/May/2004:18:54:50 +0300] "GET /****.html HTTP/1.1" 403 235 "-" "WIRE/0.1 (Linux; i686; Bot,Robot,Spider,Crawler)"
[edited by: DaveAtIFG at 7:40 pm (utc) on May 7, 2004]
[edit reason] Obscured IP [/edit]
Their bot has already downloaded over 10.000 documents from several Websites of mine, on different IPs, hosted in Greece and in US, in 6 different languages.
What's the point?
I found about WIRE bot here:
[cwr.cl...]
And I guess those people at UWashington have lots of spare bandwidth to waste...
Yep, robot programmers that don't hide normally appreciate feedback and need feedback to work on improvements and to repair bugs.
(Due to the feedback i sent to the ibm / almaden developer team a few years ago, their bot finally learned how to parse robots.txt files that are macintosh formatted - non standard cr / lf.)
Yep. The software is free and available at the Center for Web Research University Of Chile. The requests are obviously coming from the University of Washington Information School where they run several projects related to web search. The most related project is the The Web Tango Project [webtango.ischool.washington.edu], a research project including Designing interfaces for enhancing Internet searching ...