Forum Moderators: Robert Charlton & goodroi

Message Too Old, No Replies

Million Page Site - Questions

         

DocTK

6:08 pm on Nov 27, 2006 (gmt 0)

10+ Year Member



We are working on a site that will have several million pages of fairly rich content. Easiest comparison would be something like a magazine site that is providing all of its back inventory of articles and images. Each article and image will have its own page - thus millions of pages. We are doing most of the things suggested in other threads here: static text links, using sitemaps.

1. How many links on a page in the Directory?

We plan to create a Directory that will allow browsing of the content and will help search engines to spider the content. I've read the Google guidelines about 100 links or less on a page. Is that a good number to use? Do you run into issues if you have 300 or 500 links?

2. Anything else to ensure that each of the pages are indexed?

Because the content on the individual pages is pretty unique, we want to make sure that they get indexed so that while they will have low Page Rank, they still may pop because of relevance in some obscure search terms. Beyond the Directory and Sitemaps, anything else we can do to make sure they get indexed?

3. How do we roll out?

We have a domain that has an okay Page Rank where we will likely put the Directory and the Pages under this domain. However, I've heard that rolling out a bunch of pages like this all at once may hurt you. Should we consider limiting the number of pages and rolling them out incrementally? If so, what kind of roll-out would make sense?

tedster

4:14 am on Nov 30, 2006 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Hello DocTK, and welcome to the forums. I'll take a shot here, but I'd love to hear from others.

1.Do you run into issues if you have 300 or 500 links?

While Google now seems to index beyond the 100 link limit, nevertheless, the amount of Page Rank "voted" by each link to its target page goes down, divided by the total number of links on the page. It's still a good idea to limit total links.

2. Beyond the Directory and Sitemaps, anything else we can do to make sure they get indexed?

The most important is to acquire inbound links from other high quality domains. Deep links can be very helpful.

3. Should we consider limiting the number of pages and rolling them out incrementally? If so, what kind of roll-out would make sense?

On a brand new domain, releasing all at once and allowing Google to crawl naturally as it sees fit might make some sense. You will be highly dependent on getting inbound links and establishing trust in that case, but you will be in any case.

In my opinion, you are more likely to find trouble when re-adapting an exisiting domain, because this fits more closely to a spammer's profile. I would definitely consider how to roll out in a phased release. The particulars would need to be tailored to your situation - but I would suggest you start at relatively low level, and then build out more rapidly as you see a healthy response from Google.

BigDave

4:44 am on Nov 30, 2006 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Google never has had any trouble going out several hundred links. They put that 100 link guideline in because too many people were making garbage pages just for the spiders.

Build your directory in logical pieces for your users. Put as many links as make sense for those users. If you start to have pages with more than 100 links, think about how it might make sense to split it up. If it doesn't make sense to split it up, then don't.

To get the pages indexed, I will give the same advice I gave in another similar thread. It takes three things, good internal navigation, deep links and time.

Those deep links will give you multiple entry points for both visitors and googlebot. The good internal navigation (search on breadcrumbs) will put more deep pages within a click or two from your deep links.

Most importantly is to accept the fact that it will take time. If the content is that good and unique, it will start gaining those links naturally as more people find the resource. Do some moderate things to speed things along, but steer clear of trying to force googlebot to speed up.

As for the rollout, I would take a look at the sections and see if it makes sense to put out specific, more popular sections first.

If it were a medical website, diabetes would be a more popular subject than scurvy. You might have 1000 articles on diabetes.

First week, put up the diabetes section and 10 diabetes articles.

The next week put up your congestive heart failure section with 10 articles and add 10 to the diabetes section.

The third week you do strokes and add 10 articles to the previous categories.

This way you are training googlebot to accept your ever increasing amount of content and giving it time to digest it.

DocTK

5:39 pm on Nov 30, 2006 (gmt 0)

10+ Year Member



Thanks for the suggestions. Good ideas on the incremental strategy and I can see how we can do that fairly effectively.

If anyone else has thoughts - love to hear them.