Forum Moderators: goodroi
User-agent: *
Disallow: /admin/
Disallow: /download/
Disallow: /temp/
Disallow: /pub/
Disallow: /product1product_info.html
Disallow: /product1/product_info.html?manufacturers_id=6
Disallow: /product2product_info.html
Disallow: /product2/product_info.html?manufacturers_id=6
Disallow: /product3product_info.html
Disallow: /product3/product_info.html?manufacturers_id=6
Disallow: /product4product_info.html
Disallow: /product4/product_info.html?manufacturers_id=6
Disallow: /product5product_info.html
Disallow: /product5/product_info.html?manufacturers_id=6
I need to add about 250 more product Disallow pages. Is it possible to add a wild card so I don't have to enter every single one of my products pages?
say something like:
Disallow: /product*product_info.html
Disallow: /product*/product_info.html?manufacturers_id=6
Is this valid?
Thanks in advance!
thanks for the reply. Unfortunatelly, adding the metatag is not an option since these product pages are generated by the shopping cart software.
I guess I just have to enter each product one by one into the robots.txt.. :-(
Is there a size limit to robots.txt file?
---------------------------------------
Wildcards are not valid in the robots.txt standard. If you don't have 500 lines in your robots.txt file, another option would be to add the appropriate meta tag at the top of the product_info page:
<META NAME="ROBOTS" CONTENT="NOINDEX">
CHad
Disallow: /product*product_info.html
Disallow: /product*/product_info.html?manufacturers_id=6
So that
Disallow: /product
would get rid of everything following product
as in /product*/*
Googlebot does recognize wildcards
Not sure about ones in the middle of an expression though.
Besure to test with one of the validators