The org's aims are honorable, but their tools dishonor robots.txt. Every. Single. Time.
Examples from the past few days where the robots.txt they Get is always, and only --
User-agent: *
Disallow: /
-- and is immediately, and repeatedly ignored:
www.boia.org
BOIA-Scan-Agent/LC 1.0 (www.boia.org)
06:55:17 /robots.txt
06:55:18 /homepage.html
boia.org
BOIA-Scan-Agent/LC 1.0 (www.boia.org)
06:12:25 /robots.txt
06:12:26 /homepage.html
www.boia.org
LinkChecker/7.3 (+http://linkchecker.sourceforge.net/)
01/05 15:55:44 /robots.txt
01/05 15:55:45 /homepage.html
01/05 15:55:52 /robots.txt
01/05 16:09:58 /robots.txt
01/05 16:09:59 /homepage.html
01/05 16:10:00 /robots.txt
Note hits from both Hosts:
www.boia.org
= 98.174.83.170
= Mendon Cox Communications
boia.org
= 98.191.56.241
= Cumberland Cox Communications
Apparent referrers (by registered users?) are typically .edu, and also repetitive. But I seriously doubt individuals are sitting there entering my site's home page into boia.org's 'free scan' box over and over and over again at all, let alone for months on end.
Bottom Line:
Regardless of Host/IP, UA, and/or REF, robots.txt is always ignored.