Welcome to WebmasterWorld Guest from 188.8.131.52
Forum Moderators: open
joined:May 28, 2002
Just for the record, I have several dynamic URLs recently deepfreshbotted(?) Jun. 21 just like so...
including new URLs that was just created a few days before June 21.
Go figure ;)
anything after a question mark in a url, the formula being something like this
fullurl?variable1=value1&variable2=value2& etc ....
we are not just talking about getting spidered, you can get spidered all you want and wont necessarily rank for anything.
2 variable max but none will rank better every time. The more competitive the term the more important it becomes to have no vars in the url.
Tried tested and true on every size site you can imagine and as many languages. Every site that has been redone with out variable strings in the url has improved ranking, no exceptions.
If you are starting a site, always stay away from query strings. It is much easier to do at the beginning than it may be to get rid of them later.
I had a site with a static directory (similar structure to DMOZ) that got very good rankings in google - probably a combination of good structure, static pages, and page topics in the URL. I then switched to a dynamic directory because the program runs much better.
Pages with 2 variables get spidered and indexed and seem to show-up okay in the SERPS. I don't think the Google toolbar shows the correct pagerank for many of these pages though, as the entire directory shows the same pagerank. However, in the Google Directory, the pagerank is different than what the toolbar shows. The most frustrating part is that, despite getting almost all external links to change to my new site structure, the old pages often rank better in the Google SERPs.
On a side note, there was a several day period about 2 months ago where Google was spidering many more of my dynamic pages with up to 5 variables. None of those pages ever got indexed though.
How many is too many?
I'll mirror what jatar_k has stated about clean urls vs. dirty urls. Eliminate all variables, all of them. Clean and shorten the url strings as much as you can. If you use "/" to separate parts of the string, make sure to bring the most important part of the string forward, as close to the tld as possible.
You can get rather creative with your url rewriting and the url paths become very user friendly and much easier to bookmark and share. ;)
joined:May 28, 2002
How many variables in a url would cause trouble for Google? Also, what would Google consider to be a variable?
The original poster did not ask as to how a dynamic URL perform in regards to Google ranking.
By all means, I agree that static URL perform better than dynamic URL. However, there's always an exception to the rule. I can pull a few competitive key phrase with a dynamic URL at page1.
If the site is new then of course you have to use whatever means to convert your dynamic URL to seemingly static URL. Let's not forget that the main reason why Google is not really crazy about dynamic URLs, aside from session ID and spider trap, is because their crawler could cause a heavy load on the target server. Relevancy and ranking still applies.
Since this thread have move in the direction of the value of having a dynamic URL then let me ask this...
Suppose for whatever reason it's not practical to convert the dynamic URL to a static URL, is it then that this site is practically useless?
To humor the argument that dynamic URL does not rank well, what is then the impact of dynamic URLs to the overall site internal link structure?
Otherwise, if we to have to believe that dynamic URLs are totally useless then we wouldn't be using them, right?
one site lost every thing after id=1 (prod.asp?id=1&code=9999)
when you lose 100's pages you have a quick peep at you sites which use a rewrite mod and they are still standing.
g is littered with my pages that say :
We all know that most SEs have one problem or another with dynamic URLs. Google and AlltheWeb seem to have a handle on indexing dynaminc URLs that contain no more than 2 variables. I'm not 100% sure how the other majors handle them and don't really care as they don't justify the time and research to fully test the effects of dirty vs. clean URLs.
Think about your visitors. Dirty URLs are very difficult to remember. They give no indication to the visitor what the resource is other than a bunch of numbers, characters and variables that make no sense to them. There are other factors involved such as security, abstraction and maintainability.
What happens when you need to change the underlying technology on a site that has over 30,000 pages indexed? With dirty URLs, you'll have a very large task ahead of you. With clean URLs, the switch is transparent.
I strongly advise anyone thinking of building a dynamic site that they start with clean URLs from the very start. You need to spend as much time planning those URL strings as you do the structure of the site. In fact, it is mandatory as you don't ever want those strings to change. The W3C has a neat little tip when validating your site that refers to cool URIs don't change. In that tip, they discuss why your URL strings should not change for 5, 10 even 30 years.
I'm not too certain on where URL rewriting won't work. From the research and testing I've done, there has not been an instance where URL rewriting would have caused any issues other than the learning curve required to implement the feature.
You make the choice. Would you rather have...
Think about the long term effects of having dynamic URLs as opposed to clean static URLs. A new buzzword I've picked up lately is content negotiation. It's been around for a long time, but we've never really discussed it in detail. I think we'll see much more interest centered around the topic of dirty URLs vs. clean URLs.
[edited by: pageoneresults at 4:49 pm (utc) on June 24, 2003]
I think that dynamic urls developped from an innate laziness for programmers and weakness of scripting languages as they were maturing.
I can slap together a site with query strings in no time but it takes work and has to considered through out the design process to keep them out of the url and make a site dynamically static. I also know that most programmers don't know or care about search engines present company excluded of course.
Thats why all of these "great" programmers build these "great" products for use online that are completely unspiderable. Design Firms in their "great" wisdom build sites that effectively remove all of the traffic the client used to have.
Query strings are easy and passing params is common place but it is the method in which it is done that should be assessed and considered.
Best Practice? no query strings, period.
sorry, its a topic that makes me a little upset, no offence intended, I am a programmer so mostly I insulted myself ;)
Design Firms in their "great" wisdom build sites that effectively remove all of the traffic the client used to have.
jatar_k, a discussion with a design company that builds cookie cutter sites is what prompted my post in the first place.
Sell a new design
destroy all rankings and traffic
sell them seo
repeat as necessary
I was the seo and they resold it for 3 times what I charged bad, bad, bad. I only did was I was told because I had a boss but it was ugly either way.
<OT>sorry Dave, I meant REAL programmers ;)</OT>