I have a retail site targeting different countries. Unfortunately it is basically the same content but with just different prices. We have different country specific domains but they are all on one server and we are unable to implement different robots.txt. I just want to block out all the sites except so we don't get penalised for duplicate content.
So is it possible to have a robots.txt file us the following code
User-agent: * Disallow: <our UK domain>/ Disallow: <our AU domain>/ Disallow: <our CA domain>/
Or does robots.txt ignore any domain information and just look at what comes after the /. Very important we don't ruin our US rankings.
I'll bet you can easily find someone who *can* do individual files -- Good help is cheap in an economic downturn, something that "no-can-do" people should bear in mind... ;)
Use mod_rewrite or ISAPI Rewrite to internally rewrite robots.txt URL requests to different files based on the Host header sent with the client HTTP request. Or again, use a rewrite engine to pass all robots.txt requests to a PERL or PHP script which can generate different robots.txt content, again based on the Host header sent with the HTTP request. Or build this function into the script you use to generate your custom 404 error page contents, and let the robots.txt requests activate that script as well, with that script producing the robots.txt content (and a proper 200-OK server status response)... There are many ways to do it.