Forum Moderators: goodroi
Currently running Apache under Linux but have a control panel managing everything. Directories are handled with non-protected and protected directories separate under the root. (separate /httpdocs and /httpsdocs with cursory /cgi-bin, etc under each.) Needless to say it's occasionally a pain.
When creating a robot.txt should I treat secure folders and files as a subdomain? In other words do I need to have a separate robot.txt in the secure directory.
That being the case, is the statement:
User-Agent: *
Disallow: /https
irrelevant in standard web docs folder robot.txt, or does it prevent robots from *following* links in a page to the secure directory as well. (or use this in both directories?)
Related... Is it bad form to use both robot meta tag and a robot.txt? For instance on pages with a link to secure files/folders I don't want indexed, can I use a meta tag (index, nofollow) to prevent spiders crawling to that page?
Thanks much in advance.