Forum Moderators: Robert Charlton & goodroi
to the pages?
Also what else can I do to avoid some new pages being indexed by google and any other search angines?
In advance thank you for any comments. :)
To remove pages as you indicated, you need to use only one method... If you disallow in robots.txt, the page(s) disallowed will not be accessed again, so the 'noindex' meta tag will not be seen, making it ineffective.
If possible, I think the better method is to redirect the duplicated pages to a single set of pages, so you will gain the benefits of any inbound links, but if redirecting is not an option, either method you suggested should be effective, both for removing pages and keeping new pages out of the index.
Justin
Are any ways to hide just part of the pages? Such as some links or part of content, I know about "no follow", but are there other ways to hide links/content complitely and for all search engines?
But there is no html mark-up that disallows indexing for just part of a document.
There are some ways to keep portions of pages from being 'seen'. The most widely used are JavaScript and/or iFrames, but testing and determining which of these works best for you is usually up to you.
Keep in mind any type of 'hiding text' can be considered cloaking and/or spamming, so you really have to make your own determination, and use caution / discretion when implementing any system which shows different information to visitors and search engines.
I would suggest doing quite a bit of research, so you know the risk / reward prior to attempting to hide information... Also, keep in mind the way things are treated today could change tomorrow, and what was 'not seen' today, might be 'seen' as a 'red flag' in the near future.
Justin
The 'gone' pages will fall out eventually, meanwhile, just be sure that the 'new' pages are better, and therefore more likely to appear in the serps.
The fact that dead pages 'can' be found does not mean that they will (by the average searcher) - try a few keyword searches and you'll see.