Forum Moderators: Robert Charlton & goodroi
[edited by: tedster at 2:21 pm (utc) on Mar 28, 2013]
User-agent: *
User-agent: Googlebot
Allow: /
User-agent: TerrawizBot
Disallow: /
.....
disallow - The disallow directive specifies paths that must not be accessed by the designated crawlers. When no path is specified, the directive is ignored.
disallow: [path]
allow - The allow directive specifies paths that may be accessed by the designated crawlers. When no path is specified, the directive is ignored.
allow: [path]
To exclude all files except one - This is currently a bit awkward, as there is no "Allow" field. The easy way is to put all files to be disallowed into a separate directory, say "stuff", and leave the one file in the level above this directory:
The /robots.txt standard is not actively developed.
Also, you may not have blank lines in a record, as they are used to delimit multiple records.
You MUST include a blank line after each record, i.e. before the next User-agent.
you may not have blank lines in a record, as they are used to delimit multiple records.
Note the optional use of white-space (an empty line) to improve readability.
That's great for Google, what about other bots?