Forum Moderators: phranque
Is it your personal skill set you refer to, or technical limitations with regards to hosting?
It could be both. What options are available to me if I have the above scenario? Let's say that I cannot generate the metadata. Let's say that I cannot do something at the server level. What are my options?
Let's say that I cannot do something at the server level. What are my options?
Is this an academic question, or a real life problem?
There are basically two approaches in fighting the duplicate problem with dynamic websites. One approach is to rewrite all URLs to static versions without parameters and make sure that every part of the website uses only one version of the URL for a given page. For your example where the sort order of a table must be changed, cookies can be used which stores the preferred sort order per visitor.
The other approach is to accept that the site is accesible via multiple URLs, but make the site immune for it. Adding life meta tags as I do is one possibility, but you could also use Google's wildcard extension to the robots.txt. Wildcards in robots.txt are however not widely supported so that solution will only work for one search engine.
If you have neither access to the page source code, nor can add a sophisticated robots.txt, the remaining options are rather limited. Especially Googlebot has a habit of following all paths it shouldn't follow so eventually it will find duplicate URLs. Google claims that there is "almost nothing a competitor can do" to hurt a website in their index, so once Google has found the duplicates and stored them you are out of luck of removing them.