Msg#: 4517527 posted 3:03 pm on Nov 9, 2012 (gmt 0)
I use joomla for my website and automatically all those files are blocked is that good or bad, so I remove anything and if so why ?
any comments ? does anything need to be changed or is it ok ?
Msg#: 4517527 posted 7:22 am on Nov 10, 2012 (gmt 0)
Say what now?
robots.txt is for naming directories on your site that you don't want to have crawled by good robots. Bad robots don't read robots.txt -- or they do read it and head straight for the listed directories -- so you have to 403 them.
I'm guessing there are certain directories used internally by joomla that it doesn't want to have crawled, so it comes with its own robots.txt to go with its own htaccess. This presumably overwrites your own pre-existing robots.txt, so I hope you kept a backup.
Msg#: 4517527 posted 12:04 pm on Nov 11, 2012 (gmt 0)
the Disallow: directive matches URLs, left to right, that are to be excluded from crawling by "well-behaved bots". if you have any content on urls that match those paths that you want indexed then you should change your robots.txt file accordingly. however excluding a URL from being crawled does not prevent the URL from being indexed, it prevents the content from being indexed. if you don't want either to appear in the index you must allow crawling of the URL and provide a noindex signal such as the meta robots noindex element for HTML documents or the X-Robots-Tag HTTP Response header for other content types.