Forum Moderators: open
Thanks
If your cart has long URLs - especially with lots of vars - then some bots will choke/refuse to follow them.
You might want to read this:
[webmasterworld.com...]
would end up like this:
site.com/23/1/234
I beleive there is an old article on it over at sitepoint.com if you try their sitesearch for "search engine friendly urls" or similar...
Welcome to WebmasterWorld!
Nick
Let me see if I understand, because I've seen the same with google. It does spider my PHP pages from the static HTMLs however it doesn't follow any links from the dynamic pages. It does index the first dynamic page, but then it stops :(
So instead of having:
mysite.com/myprog.php?cat=235&startitem=8 ( for the next page )
You should have a link like:
mysite.com/myprog/235/8
And then it will be indexed. Because I do have a lot of page which still haven't been visited. The only thing I have to wonder about is how to pull it off from a programmers point of view...
I always thought it didn't recurse the dynamic page to avoid picking up double content..
It's pretty and very, very, simple. The only thing not mentioned was that I need to add a base href=mysite.com in the head tags because I got broken image links, but it works splendidly :) :) Let's see if it will be crawled... (Can't wait to check the logs tomorrow...)
Thanks a lot :)
I don't use php etc, but remember reading somewhere here that Google only crawls deep into a dynamic site with long url strings if the pr of the index page is pr5+.
However they have been trying to crawl more dynamic pages recently and certainly the serps are showing more of these in the top 10.
In the meantime the above advice is definately the way forward and works.
However there are hardly any calls to PHP programs from the index page; just static pages. Those pages do have links to PHP for further browsing. Maybe it doesn't advance, because those pages are usually PR3-PR5.
Anyway it doesn't matter anymore, because I do like this idea. It makes for easy to read URLs and that's important as well.
With regards to the base URL. I (still and always will) use Netscape Composer to create the pages because it's easy, works on all browsers (unlike some products...) and generates clean code which can easily processed in program files.
Googlebot and Scooter did do quite a lot of crawling last night, but didn't go to the PHPs yet, because I only implemented it on a couple of pages, so let's wait and see...