Forum Moderators: phranque
Example
/images/
/images1/
/bluewidgets/
How can i stop this from happening?
If say i was to disallow them in the robots.txt file, wont it stop the .html files inside the directory from being indexed?
Thanks
Yes it would, because robots.txt uses prefix-matching; If the to-be-requested URL prefix matches the prefix specified in a Disallow directive, then that URL won't be fetched.
The usual solution on Apache is to use
Options -Indexes
However, the question remains: How did Google find URLs pointing to your directories? I am seeing Yahoo! Slurp attempting to spider unlinked directories recently, but I don't recall Google ever attempting to fetch unlinked URLs (with the obvious exceptions of robots.txt and sitemap.xml).
Jim