- - Search Engines
- -- Sitemaps, Meta Data, and robots.txt
- ---- HTTP 320 URLs, robots.txt and Google
nabucosound - 7:43 pm on Jun 15, 2012 (gmt 0)
I have a homepage listing a bunch of items. Each one has some information about a particular external page (image, title, description, some other info bits).
Each item also has a link to a HTTP 302 Redirect. For example: /go/122456/
On that intermediate redirect step I capture info about the user/request (such as datetime, referrer, browser, etc).
Redirect goes to the final page, always an external site.
Google has now indexed my redirect URLs in a way that it shows SERPs with the title and description from the final external pages, but with my redirect URL (again, /go/123456/)
This was reported as a bad practice by a SEO tool I am using. I want now to remove those URLs, before they hurt my site.
I added a robots.txt to the root of my website with this information:
Google webmaster tools seems now to follow the new restriction, as per the number of "Blocked URLs" on its "Blocked URLs" under "Health" section.
URLs continue to show up on SERPs, when I do a search like "site:mysite.com"
Shouldn't Google remove automatically previously indexed URLs? Or may I do it by hand ("Remove URLs" section, under "Optimization")?
Thanks in advance,
Thread source: http://www.webmasterworld.com/robots_txt/4466000.htm
Brought to you by WebmasterWorld: http://www.webmasterworld.com