Well, let see if I may demo a case where using Pascal Casing in a full URL
I am building as a section of a new site some form of “…pedia”
Here I could not care less if I use Pascal Casing and if the whole url is case sensible.
Why: at three levels deep, even four when adding the whole www.DN/
No one will type it, in order to get a page coming from four levels deep
The page will only be accessed via site nav or indexed by SE
In both case no one will hard type its full URL
So I believe that in this case case-sensibility is not an issue
Actually, does anybody try to memorize a full URL?
Did you every try to memorize this or just www.domain.com?
If I find an interesting page / story, I simply bookmark it.
The upside to example.com/PascalCasing.html is that some stupid bots convert the URLs they scrape to lower case and then all they get is 404s.
It's a blessing ;)
Can Google have different mixed case of the same URL in its SERPS? Might not be good for PR, and might not be good for any adsense either - as Google could be thinking the ad is being displayed on a different page that gets very low impressions.
Should I be redirecting incorrectly cased URLs to a lowercase URL?
|Should I be redirecting incorrectly cased URLs to a lowercase URL? |
Yes. Or, in this instance, to their "ProperCase" if applicable. ;)
This type of casing causes duplicate content issues. What happens is that some Web masters only use lower case letters in links. Therefore when they link to you, it creates two unique pages in Google for the same content: www.domain.com/file.htm and www.domain.com/File.htm . These are seen as separate pages by the engines and their Page Rank is not consolidated unless Google figures out they are the same page (which is A LOT better since Big Daddy but still not perfect). And since there is no proven benefit to having letters in a URL string capitalized, it makes no sense to do it.
Remember, we are dealing with two distinct areas here; the domain itself and then the file naming conventions afterwards.
If you have the proper facilities in place for redirection, whether you're using PascalCasing in the domain itself and/or in the file naming conventions, everything should be just fine.
Let's drop the "file naming" aspect as that is where all of the technical issues come into play. Doing www.webmasterworld.com or www.WebmasterWorld.com doesn't matter. Domain names are not case sensitive. File naming is.
Sorry to join this discussion so late..
Not mentioning any names but 'someone' in our office recently made some changes in our back end CMS and, as a result, around 30% or the URLs to our site have changed with regards to the casing. For example; www.domain.com/page.htm has changed to www.domain.com/Page.htm.
Both of these URLs work but I'm wondering what the best way to handle the situation is as most of the URLs in the Google index are the old versions and Google could possibly see all of these pages now as being duplicates.
|Both of these URLs work but I'm wondering what the best way to handle the situation is as most of the URLs in the Google index are the old versions and Google could possibly see all of these pages now as being duplicates. |
You are correct. Depending on your platform, and whether or not you have the capabilities to do so, you'll need to make sure that one or the other is "forced". The recommended would be lower case for file names as they are case sensitive. You can force the Pascal Casing for "file naming conventions" but I do believe you'll add a "challenging layer" in your indexing process.
|Not mentioning any names but 'someone' in our office recently made some changes in our back end CMS and, as a result, around 30% or the URLs to our site have changed with regards to the casing. |
If those are getting indexed right now, you'll need to make sure that a 301 is being returned once you force the case.
Take it from someone who made the choice to mix case in URL's a long time ago -- DON'T !
Due to my legacy webserver, I wasn't able to do a Mod ReWrite to fix the problem, and the only way I can address it is via ROBOTS.TXT Disallow control. That is, I specifically disallow all lowercase URL entries for effected sites.
Failure to control it, caused me duplicate content headaches, which I now have under control.
But, if I had to do it all over again - I would have stuck with an all lowercase URL naming convention.
| This 40 message thread spans 2 pages: < < 40 ( 1  ) |