Forum Moderators: open
I encouraged them to join the discussion here.
Also, I said it's mostly because of three issues. Being hosted by HE. Not obeying robots.txt. Not being the kind of site that will send most of us any meaningful traffic in exchange for letting our sites be crawled.
2) Most of us don't consider whether the bot-owner 'cares' or not, we just don't want them wasting our bandwidth or scraping our sites -- intentionally or unintentionally outranking us in the SERPs with our own content.
3) This isn't us 'being mean' to them, it's just self-defense -- especially in cases where the Webmaster is on a limited-bandwidth plan, and the aggregate 'junk-robot' traffic puts the site over that limit and incurs an additonal hosting charge -- with no benefit to the site whatsoever.
4) In many cases, it's simply a matter of the Webmaster not having time to pursue DMCAs and other actions that would otherwise be required to 'keep up' with all of these unknown, useless, and/or malicious requests. It's far easier to just allow legitimate browsers and the top-tier search engines, block the rest, and go on with other more-productive site development activities.
5) To each his own. Each Webmaster is free to protect his/her work and intellectual property if he/she chooses to do so. Or not.
Jim