Forum Moderators: Robert Charlton & goodroi
I have identified a range of key areas that need to be tackled that are creating duplicate content issues.
After tidying up the robots.txt file to prevent further indexing of duplicate (dynamic) pages, and placing the necessary redirects on-site to so that duplicate pages eventually fall out of the index, I am looking at direct removal requests to speed things up.
I have had some success with the directory exclusion, but I have a question now we have come to the dynamic URL removal.
On the directory removal request G. seems to work well with the hierarchical URLs, but how would it treat a dynamic URL such as:
www.example.com/index.php?dnurl=various&dynmaic=parameter+followed-by-old-session-IDs
If the following was submitted for directory exclusion, would all URL's with that prefix be de-listed, or would G. look at the last directory and delist everything to the root.
Exclude: www.example.com/index.php?dnurl=various&dynmaic=parameter
[edited by: Receptional_Andy at 11:34 am (utc) on Feb. 10, 2009]
[edit reason] Please use example.com - it can never be owned [/edit]
A dynamic url is not the same thing as a directory. If you request removal of a dynamic url then that exact url is what gets removed - just the version with those parameters.
What I would do in your situation is just use a robots.txt disallow rule. Google will then stop spidering those urls and eventually they will fall out of the index.