Forum Moderators: Robert Charlton & goodroi
Crawl errors is one of the most popular features in Webmaster Tools, and today we’re rolling out some very significant enhancements that will make it even more useful.
We now detect and report many new types of errors. To help make sense of the new data, we’ve split the errors into two parts: site errors and URL errors.
Look up the phrase "soft 404". Google hates 'em.
Not found nn
Not followed nn
Access denied nn
Server error nn
Soft 404 nn
Other nn
vanessafox wrote:
Access denied appears to primarily be URLs that returned a 401 status.
403s are listed in the "other" category.
[edited by: tedster at 3:43 pm (utc) on Mar 13, 2012]
I have questions into Google asking about the removed functionality (particularly the confusing changes such as the Not Followed errors) and I’ll update this story as I hear back.
Previously, you could download up to 100,000 URLs with each type of error. Now, both the display and download are limited to 1,000. Google says "less is more' and "there was no realistic way to view all 100,000 errors - no way to sort, search, or mark your progress." Google is wrong.
Every page that I have ever made and removed, going back 7 years. They have dug up EVERYTHING, as if everything I ever did should still be on the web.
many of the URLs in the "blocked by robots" list aren't actually errors, the webmaster purposely blocked that access