Forum Moderators: Robert Charlton & goodroi
In any case, there are a few ways to handle. Perhaps these are duplicates that you don't want/need indexed? Like the print versions of article pages or something?
If you have links on your site that point to duplicate pages or pages which are undesirable to have indexed, you might place the rel="NOFOLLOW" attribute within their A HREF tags.
Also, you could set up their META Robots tag to specify that the page should be NOINDEXed.
Finally, if they all occupy the same subdirectory on your server, you could specify that they not be indexed in the site's robot.txt file.
At the moment they appear to be orphan pages and my guess is that they were linked to at some time in the past on a previous incarnation of the site.
The problem is the best way to get rid of them? Either remove them from the index by using the methods you outlined or just by simply deleting them from the site.
The robot.txt sounds like the best idea because they’re all in roughly the same directories. Do you think removing lot’s of pages in one go is a good idea or should I try and phase the removal in gradually?