Forum Moderators: Robert Charlton & goodroi
So in order to prevent the world at large from knowing that a given page exists, you have to let google run rampant over the entire page, trusting it to obey your "noindex" wishes? This is not a pretty choice.
If you want the page completely out of the SERPs, you need to not disallow the page in robots.txt and then use the meta robots noindex directive on the page itself.
Robots.txt is the suggested method of handling redirect pages, any suggestion on getting those to remain out of the index?
If you wish to remove your content using the URL removal request tool in our Google Webmaster Tools, you must first meet the criteria listed below.
To remove a page or image, you must do one of the following:
* Make sure the content is no longer live on the web. Requests for the page must return an HTTP 404 (not found) or 410 status code.
* Block the content using a robots.txt file.
* Block the content using a meta noindex tag.
To remove a directory and its contents, or your whole site, you must ensure that the pages you want to remove have been blocked using a robots.txt file. Returning a 404 isn't enough, because it's possible for a directory to return a 404 status code, but still serve out files underneath it. Using robots.txt to block a directory ensures that all of its children are disallowed as well.