homepage Welcome to WebmasterWorld Guest from
register, free tools, login, search, pro membership, help, library, announcements, recent posts, open posts,
Pubcon Platinum Sponsor 2014
Home / Forums Index / Search Engines / Sitemaps, Meta Data, and robots.txt
Forum Library, Charter, Moderators: goodroi

Sitemaps, Meta Data, and robots.txt Forum

Multiple Directories
robot.txt, subdirectories, protected directories

10+ Year Member

Msg#: 559 posted 1:47 pm on Feb 28, 2005 (gmt 0)

Greetings all. My question regards multiple directories under a root outside of the main web docs folder...

Currently running Apache under Linux but have a control panel managing everything. Directories are handled with non-protected and protected directories separate under the root. (separate /httpdocs and /httpsdocs with cursory /cgi-bin, etc under each.) Needless to say it's occasionally a pain.

When creating a robot.txt should I treat secure folders and files as a subdomain? In other words do I need to have a separate robot.txt in the secure directory.

That being the case, is the statement:

User-Agent: *
Disallow: /https

irrelevant in standard web docs folder robot.txt, or does it prevent robots from *following* links in a page to the secure directory as well. (or use this in both directories?)

Related... Is it bad form to use both robot meta tag and a robot.txt? For instance on pages with a link to secure files/folders I don't want indexed, can I use a meta tag (index, nofollow) to prevent spiders crawling to that page?

Thanks much in advance.


Global Options:
 top home search open messages active posts  

Home / Forums Index / Search Engines / Sitemaps, Meta Data, and robots.txt
rss feed

All trademarks and copyrights held by respective owners. Member comments are owned by the poster.
Home ¦ Free Tools ¦ Terms of Service ¦ Privacy Policy ¦ Report Problem ¦ About ¦ Library ¦ Newsletter
WebmasterWorld is a Developer Shed Community owned by Jim Boykin.
© Webmaster World 1996-2014 all rights reserved