Forum Moderators: goodroi
Can anyone suggest a good way of doing it?
Right now, we're thinking about denying access to all pages and then granting access to the home page.
User-agent: *
Disallow: /
Allow: /index.html
[en.wikipedia.org...]
as Google, Yahoo, and I think Ask support the Allow function (not MSN?). But then we have the issue of how to get the home page indexed for a call to www.widgets.com. Is it possible to 301 permanent re-direct calls to www.widgets.com to www.widgets.com/index.html?
Is there a better way that isn't as icky?
Thanks,
Steve
User-agent: *
Disallow: /*/
Apparently works in Google Webmaster Tools robots.txt validation. Google Webmaster Tools robots checker says that www.example.com/ is ok but will not crawl www.example.com/blah. I think Yahoo will allow this given their documentation at
[help.yahoo.com...]
One problem is that we haven't been able to figure out if MSN allows this or not. The other problem is that technically speaking the robots.txt formal specification does not support wild-carding in the disallow although Google and Yahoo support it. :/
I guess we'll go back to denying specific directories instead. Anybody have any experience with /*/ in Disallow?
Steve