Forum Moderators: goodroi
This morning I've been reading some online tutorials but cannot find the answer to this question, so thought I'd better post it here:
Is there a way to indicate with one line of text that ALL sub-folders are to be disallowed? The key here is to not block the bot from indexing the top level html pages, as that is where the important content resides.
For example, would this work:
Disallow: /*/
Or does that block all the top level html files as well? I realize I can individually list each and every folder, but am hoping for a more efficient solution.
As always, any advice is appreciated...
ps. Am sorry to say that after my initial success with msnbot to only 4 MB bandwidth a day (down from 40+), they are now up to 14MB/day, with very little new content added to the site during this time period. Very perplexing.
.......................................
A really good way to see how robots.txt files are done properly is to check major sites' files. E.g.:
eBay [ebay.com]
craigslist [craigslist.com]
And here's a whopper:
Google [google.com]
And last but not least, a blogger!:)
WebmasterWorld [webmasterworld.com]
(actual [webmasterworld.com]; original [webmasterworld.com])
.
P.S. to Newcomers
The Web Robots FAQ [robotstxt.org]