Forum Moderators: goodroi
Yes exactly robots.txt blocking pages i didn't want blocked
i also include simple robots.txt file but can't over come those 58 blocked url's.
didn't gave him a permission to blocked
To be honest, I am not sure why would you want to allow Google to crawl these pages anyway, especially /wp-admin/.
iam using simple robots.txt file, but failed
User-agent: *
Disallow: /wp-admin/
Disallow: /wp-includes/
Sitemap: http://www.example.com/sitemap.xml.gz
User-agent: *
Disallow:
Disallow: /wp-admin/
Disallow: /wp-includes/
Sitemap: http://www.example.com/sitemap.xml.gz
It looks like the addition of a Disallow:
line might address the issue if that's all there is in the file. There is nothing allowed, only disallowed.
[edited by: aakk9999 at 1:20 pm (utc) on Nov 16, 2013]
User-agent: Googlebot
Disallow: /*?
Disallow: /*newwindow=true
Disallow: /*dur=124
Disallow: /*dur=0
Disallow: /*replytocom
Disallow: /*refresh=1
Allow: /This is not standard. Instead of this, you should have:
# Google Image
#User-agent: Googlebot-Image
#Allow: /*
Allow: /wp-content/uploads/
# Allow Google Image bot on entire site
User-agent: Googlebot-Image
Disallow:
# Allow Google AdSense bot on entire site
User-agent: Mediapartners-Google
Disallow:
# Allow Google Adwords bot on entire site
User-agent: Adsbot-Google
Disallow:
# Allow Googlebot for mobile on entire site
User-agent: Googlebot-Mobile
Disallow:
# Googlebot - Disallow all URLs with query string parameters, allow the rest
User-agent: Googlebot
Disallow: /*?
# Any other bot not listed above - disallow wp-admin and wp-include directories
User-agent: *
Disallow: /wp-admin/
Disallow: /wp-includes/
Sitemap: http://mysiteurl/sitemap.xml.gz
Reading your latest post, I am not sure what is the problem exactly?
How your directories are organised on the server is irrelevant, what is important is what you see when you request robots.txt or sitemap.xml which should be in your domain root of each of your domains.