Forum Moderators: open

Message Too Old, No Replies

Insufficient Spidering

Spider not searching all pages

         

Valhalla

5:44 pm on Mar 21, 2003 (gmt 0)

10+ Year Member



Google Spider is only creeping around 1800 pages, instead of the full 6500 of our site. This is not a typical site, but more of a library and therefore it is imperative the spider successfully visits each page. Any idea as to why it stops before completing it‚s task?

Dolemite

5:47 pm on Mar 21, 2003 (gmt 0)

10+ Year Member



That's quite a lot of pages. How long have these pages been up?

Provided the googlebot has a reasonable path of links to get to each of these pages, they should eventually be spidered.

Mohamed_E

6:35 pm on Mar 21, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I believe that the maximum number of pages Google will spider on a site depends on its PR. Makes sense: if a site is "important" spider it in depth, if it is less so spider it less fully.

vitaplease

6:17 am on Mar 22, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Try to get some external deeplinks into your more inner pages.
Try to make "your last page" as few clicks away from your index page, whilst limiting the number of links on each page.

see also:
[searchengineworld.com...]

new_shoes

12:15 pm on Mar 22, 2003 (gmt 0)

10+ Year Member



Are your pages created dynamically? If a session id appears in the url, it has known to give problems in the past.

I went from

site.com/file/article.php?id=123&session_id=AAREW42432

to

site.com/article/123-article-about-widgets

The "-article-about-widgets" string is just fill, only used for google - when a user views the file "article", which is actually a php file, everything after the first "-" is stripped, so I get the article id.

It's easy to implement, sticky me for instructions if you need them.