Forum Moderators: goodroi
User-agent: Googlebot
Allow: /
User-agent: bingbot
Allow: /
User-agent: Twitterbot
Allow: /
User-agent: YandexBot
Allow: /
User-agent: *
Disallow: /
there may be no need to have dozens of specific bot disallows entries in one's robot.txt fileThere's also something else to consider... if you block the other bots by default, you may not get their actual bot w/ the complete UA to research.
if you block the other bots by default, you may not get their actual bot w/ the complete UA to researchOh, that's a point. Another is that requests for robots.txt may not contain exactly the same headers as requests for “real” files, so you end up poking the wrong holes, or more holes than is strictly necessary. It probably isn't specific to robots.txt; I suspect you'd see the same variation in requests for any .txt file. As with the UA*, this is rare but it does happen.