Forum Moderators: goodroi
Is there any limit for the robots.txt file, I was thinking about creating approximately 600 unique rules to avoid spiders viewing some of my pages. Would this be a problem? I know the best way is to use a wildcard but in this case it can be used for certain resaons. What would be the problems that can arise if I use a robot.txt file with 600 or more rules (no idea how kb will this be)?
Any ideas will be appreciated, thank you in advance
Instead, you should look at changing the URL of everything you don't want crawled into a sub-directory and just block the directory with robots.txt. It sounds like that might be a lot of work on your end; but if you want the rest of your site crawled that's what I think you've got to do.
I could be wrong of course; but if I was googlebot there would without doubt be an upper limit to how much robots.txt processing I am prepeared to do in order to crawl your domain.