| 12:32 pm on Jul 5, 2005 (gmt 0)|
Google and other SEs have no issues with session IDs. They just don't crawl them a lot, because their Web robots would end in an infinite loop if they would try to follow every link. The URL is unique including the query string. Since usually a bot gets a new session ID assigned per visit, it would index the same content over and over from different URLs.
Site owners do have issues with session IDs. It's good practice to detect spiders, serving them a clean URL without long and ugly ID variables. However, all page content should be identical, regardless whether the script is called with a session ID or not.
Geeks have no issues with session IDs. They love session IDs. Unfortunately, they aren't involved in marketing tasks. Otherwise they would apply a more elegant Web development.
| 5:56 pm on Jul 5, 2005 (gmt 0)|
At the conference a Google engineer re-affirmed that we should let Googlebot crawl the site without session ID or Cookie.
| 8:07 pm on Jul 5, 2005 (gmt 0)|
There are work arounds to get the best of both worlds (having a session ID and having a clean URL) without needing to detect the user agent or IP
| 9:30 pm on Jul 5, 2005 (gmt 0)|
I could be wrong but I donít think you understand what a session id is when you start talking about database stuff.
This is NOT a session ID and Google index's them fine after the first page is delayed:
Its a natural way of providing SE with how your site can work rather than mod re-write. Mod re-write can be a pain for advanced sites and bug tracking and impossible for some ASP servers.
These however can be session id's and are bad
a more typical session id looks like this:
never pars you session ID's via a URL! In ASP its easy... dunno about PHP.
Google guidlines suggest this too could be bad:
I agree it should be written like this:
Hope it helps.
| 4:26 am on Jul 6, 2005 (gmt 0)|
I have seen some websites with session ids and still indexed. I would recommend though configuring webservers to be session transition capable. That technique worked well with me.
| 11:46 am on Jul 6, 2005 (gmt 0)|
[There are work arounds to get the best of both worlds (having a session ID and having a clean URL) without needing to detect the user agent or IP]
Can you tell us more about these work arounds?
| 11:47 am on Jul 6, 2005 (gmt 0)|
[I would recommend though configuring webservers to be session transition capable. That technique worked well with me. ]
Is there someplace we can get more info on configuring webservers to be session transition capable?
| 2:49 pm on Jul 6, 2005 (gmt 0)|
Most Session IDs in URLS are 'caused' by blocking cookies.
|Allow search bots to crawl your sites without session IDs or arguments that track their path through the site. These techniques are useful for tracking individual user behavior, but the access pattern of bots is entirely different. Using these techniques may result in incomplete indexing of your site, as bots may not be able to eliminate URLs that look different but actually point to the same page. |
And GoogleGuy has also warned - going back years - that session IDs in URL strings are bad news.....
| 3:13 pm on Jul 6, 2005 (gmt 0)|
|Can you tell us more about these work arounds? |
1. Save that information in a cookie and not a querystring.
2. Make your site work, even if the browser does not support cookies. Maybe assigning a kind of default session ID for the browsers that do not support cookies (like most SE bots)
| 4:26 pm on Jul 6, 2005 (gmt 0)|
Try an Google Search to find out howto in PHP. In ASP it holds session data in memory I think.