Forum Moderators: open

Message Too Old, No Replies

Can a Load of New Pages Hurt an Existing Site?

Not sure I've Heard Comments on This

         

caveman

11:47 pm on Sep 27, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



One of our mid sized sites became quite a bit larger two weeks ago, when we added about 500 pages of content...a section we had wanted to add for a while now. The new section provides greater detail and info on a subtopic of our site which was previously only covered by a single page.

Anyway, the site just took a major hit across all pages, and it's a rare time when we've wondered if we were affected by the algo element(s) that some refer to sandboxing.

Any experience along these lines? We were not aware of an existing site being hurt *just* by virture of adding 30% more pages. (No dup issues or anything obviously like that at play here.)

caveman

6:39 pm on Sep 29, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Midhurst, I already answered the question, above:

There are text blurbs for each entry, and we write every one of the entries ourselves; 10 or 20 entries per page...thousands of entries across the site. And there are also contributed articles (hundreds) from enthusiasts with interest/expertise. ...

Also, I used widgets to illustrate, but the site is not about things for sale. Mainly information. It's closer to a site about butterflies than it is to a site selling books or bbq's. ...

IF the site were about butterflies, which it's not, and we added a new section on European varieties, when the site had mainly covered N. American varieties, one might expect the new pages to be structured similarly to the old, as is the case here. ...

...all we did was to expand a previously meager section of the site (single page) to be comparable in quality and structure to the other sections. Doing it in pieces would have made the new expanded section look odd/incomplete, so we waited until it worked well as a stand alone entity.

All original content, as noted. Uploaded when the new section was complete and ready for prime time.

onebaldguy

6:54 pm on Sep 29, 2004 (gmt 0)

10+ Year Member



Everyone in this thread is talking about a possible penalization for an increase in pages. We did the opposite. We had a lot of old pages that we wanted out of the index and so we cut our number of pages in about half (by putting a no-index tag on the older pages) and the # of pages in the google index was reflected in google in August. We did this to ensure we would only have good content pages in the index.

Then we saw a dramatic drop on Sept 9. The pages that were dropped out were deep pages that did not have much PR, so I don't think we were hurt because of a PR loss, we have some great links coming in to our important pages.

We also noticed we stopped getting fresh tags about 3-4 weeks after the # of pages were cut in half.

caveman

2:23 am on Oct 3, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Just an update. The site has not come back. We continue to suspect that the addition of too many pages too fast caused the entire site a bit of trouble. No proof as usual, but the addition of this new section was the only notable change prior to the site's banishment. It would seem that the addtition of bulk pages in some way played a key role.

My best guess is that we suddenly appeared 'abnormal' ... but in this case, a 'best guess' and a 'good guess' may not be the same thing.

Jane_Doe

6:57 am on Oct 3, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Did you try restoring the site to the old version prior to the new pages being added?

guynouk

2:04 pm on Oct 4, 2004 (gmt 0)

10+ Year Member



Caveman, do any of these additional pages have common external links on them?

caveman

4:10 pm on Oct 4, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



We have not yet rolled the site back ... we're trying to see if we can revive it with some tweaks (I always prefer to move forward rather than backwards if possible, but we're keeping the roll back as an option, certainly).

And no, there were not common external links to the new pages, but that was a great question. :-)

Go60Guy

4:26 pm on Oct 4, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Have you studied this thread? [webmasterworld.com...]

It might provide some insights.

caveman

5:03 pm on Oct 4, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Go60Guy, yes in fact I made a few posts in that one.

I'm pretty conviced that G has been tweaking their dup filters on any number of levels: internal and external links, anchor text, body text, page structure/templates, etc.

We're trying to sort out now what aspect of the new pages may have caused this. We're looking closely at links.

lufc1955

5:16 pm on Oct 27, 2004 (gmt 0)

10+ Year Member



Wish I had read this thread before I increased the size of our site by 50%. Shortly after doing this our site was dropped by Google completely. Internal linking was not a problem and only very minor other changes were made. I am convinced that adding too many pages at once is now a big problem. Has anyone who suffered this penalty seen any improvement in their situation?

wiskur

5:45 pm on Oct 27, 2004 (gmt 0)

10+ Year Member



Same thing happened to us recently. We currently have about 1500 pages. I had a chance to make about 50 new pages and uploaded them about 2-3 weeks ago.
I noticed our sales have dropped the last week or so. I just ran report and we have dropped for most of our terms in Google (some quite sharply). Nothing else had been changed on any existing pages. These new pages all had internal links with no external links.

caveman

8:49 pm on Oct 27, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Just got a couple of emails about this so I'll update. Unfortunately we have not seen the site come back yet. We run enough sites that we have a decent understanding of what tends to get sites punished by G, and have never been able to find anything else wrong. (Plus the only change we made to the site around the time we got hurt was the addition of those pages.)

However, I can't prove that the new pages had anything to do with the problems.

I'd be curious if many other members could cite recent situations where they added a substantial number of new pages to a small/mid-sized site and saw their entire site drop in rankings.

Also, more important, are there many who can say they've added a large percentage of pages with no adverse affect? I frankly still keep thinking it has to be something else. If *new* sites have trouble ranking, and *old* sites can't add pages, then G would essentially be strangling the Web. :-(

Robber

9:09 pm on Oct 27, 2004 (gmt 0)

10+ Year Member



We had an interesting one, one of our sites had been around for a few years, sitting at PR5/6 with loads of top spots etc, it was about 10000 pages in size.

The trouble was, not only had it been around for years, but it hadn't really been updated for a few years except for adding the extra bit of content to pages here and there. We wanted to start getting more out of the site in terms of revenue so it needed an overhaul.

We knew of the suspicions that changing a site could lead to problems so we were careful. Every single page (read exact URI) from the original site was maintained including main category, sub category, content and site map pages. They were left there and the linking structure was kept identical.

Regarding on page factors the template is basically the same with some small adaptations but the bulk of the facelift was done by tweaking the external CSS.

The biggest changes were:
1) to add a significant amount of new content to each page
2) we launched 2 or 3 new categories, about 180 pages per category

So, where are we now? Well basically Google dumped all of the old pages in mid September, despite our efforts to keep them looking pretty much the same in terms of their underlying code.

Whats is funny though, the new categories actually rank prety well, in fact really well in some cases - where I would have expected them to rank if they had existed on the old version before the changes.

Somethig else, we thought we might be on the come back trail last week. One of the search term where we used to be #1 got dumped to about #90, but every now and then we were seeing it it pop back up to #1 (and #2 if you count the indented result). But this was really intermittent, even when querying just the one data center where it was showing. Sadly though, this week we don't seem to be figuring at all. I am pretty confident that these pages will surface again though!

I must admit I don't really know why this could be but its probably something along these lines:
* many pages updated at once
* google somehow thinks their is an old version and a new version which are duplicates, but I can't see how as all the URLs are the same
* the site was going to get hit anyway and its just a coincidence
* because the pages are now new they must meet some additional criteria (but then why would the new pages that didn't previously exist rank well)
* because the pages have been updated G thinks the links on those pages are also new now and must mature before target pages get benefit (again though, what about the new pages)
* etc etc...

Basically more questions than answers at this point!

randle

9:41 pm on Oct 27, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



We have suffered negative consequences from adding new pages to our sites. The amount of pages has been low, but the sites we have are not that large. They average around 30 pages and we added approximately 3 pages, making it a 10% increase in size.

What we have found is that any new page comes out with a PR of 0 causing the site to now be linking out to 10% of its total, to pages that have a PR of 0. In this situation we did not drop off the map, but dipped considerable about 20 places. What complicates the matter is the sandbox, now those new sites are stuck with PR 0 for a long time. The recent PR update did get the bars to go green on about 20 new pages we had, and alas most sites moved back up.

What troubles us is we want to expand, and provide new content for our visitors, but in this climate the only way we can do it and maintain our traffic, is to add way slower than we would like to. If we added 50% new pages at once, I am convinced we would fall at least 100 spots in the SERP’s for our main key words. I know this doesn’t help explain the problem, but at least you’re not alone!

CodeJockey

12:54 am on Oct 28, 2004 (gmt 0)

10+ Year Member



Our experience is a variation of Randle's...

We split off four highly specific topics to their own domains roughly four months ago. (All of the domains are still showing PR0 for every page.)

After that, we added quite a few new pages (probably 15% or so in terms of total pages) to the remaining topics. We don't do anything weird, and the vast majority of the pages are content pages with only Adsense on them. We've basically disappeared from Google, but interestingly enough, almost all of the other search engines think that the content is useful and rank us fairly decently.

The only thing that we think could possibly explain the disappearance is that new pages are allocated a PR0 and a substantial increase in PR0 pages must trigger some sort of lowering in importance of the index page, even though the Public Page Rank shows as PR5/6 on all of the disappeared pages.

northweb

1:18 am on Oct 28, 2004 (gmt 0)

10+ Year Member



We have increased our site by 20% - html pages and have not seen any drop in ranking

Vec_One

2:50 am on Oct 28, 2004 (gmt 0)

10+ Year Member



I think randle is just about right on the money. New outbound links start losing PR immediately. The pages being linked to, however, receive little or no PR for a period of time.

I don't have any proof of this, of course, but it would explain the phenomenon perfectly.

This might actually be an improvement to the algo. It would reduce the effectiveness of link selling, blog spamming and throw-away domains, as well as many affiliate sites. The BBC's of the world would be largely unaffected because their growth is usually not extreme in relation to their site size.

God/Google help anyone who makes a major change or addition to a site though.

caveman

4:50 am on Oct 28, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



If this is happening on a large scale - and we're far from knowing that it is - any thoughts on whether it's new *pages* or new *links* that are being discounted? I've generally felt for a while now the the CW saying it's links is not necessarily accurate... :-/

grandpa

5:32 am on Oct 28, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Very interesting reading. It's suggesting that one part of the algo determines a percentage of growth on a site, and there may a triggering at a specified percentage or number of pages added.

What if the new pages were to be added with the meta directive robots - NOINDEX. I know this defeats the purpose for adding pages... but would that defeat the problem? Speculating that would actually work, the directive could then be selectively changed over a period of time.

Vec_One

6:03 am on Oct 28, 2004 (gmt 0)

10+ Year Member



That's right grandpa, I think northweb might have been just under the radar by a few % points. I wish I had northweb's luck!

The NOINDEX seems like a good idea for some situations.

BTW, my site grew considerably, but much of the growth was months before I got hit on Sept 22. I suspect that it might take quite a few months for a page/link to mature.

neuron

6:20 am on Oct 28, 2004 (gmt 0)

10+ Year Member



I must've had one of the first sites to go into the sandbox from adding more pages. About the first week of last March I published about 600 more pages on a 200-page site. To that point the site was ranked remarkably well, and still is at MSN and Yahoo. Between 6 and 8 weeks after the new pages went live, the site vanished for any decent placement in the SERPs. It was still in the index, it just couldn't pull any more numbers between 1 and 3, let alone under 50.

Now I think I've done it again. I had a site I hadn't touched in over a year because of some issues that have nothing to do with SEO or websites for that matter. It was still ranking decently for some keywords in Google, but after reading all the sandboxing discussions I realized this site was not in the sandbox and thought I'd give it a push and see what happens. The site was 40 to 50 pages and I added about 80 dynamic pages to it, all should be spiderable and indexable. I expect that will push it into the sandbox. Here's why I think that:

I'm not going to go into the numbers or the various theories themselves, but I assume the main index is full and that's why there is a supplemental index. It is not pages that are put in the main index but sites. New sites go into the supplemental index. Sites in the main index are going to grow. The main index can only hold a fixed number of pages. Thus, as sites in the main index grow, some must be moved to the supplemental index, since both the main and supplemental indexes hold whole sites, that is, you won't find half a site in the supplemental and half in the main, at least not for more than a few weeks. As sites in the main index grow, which ones are selected for retention and which for migration to the supplemental? Sites in the main index that have the most listings in the SERPs will be selected to remain the in the main index over those that do not have as many SERPs listings. Sites that do not grow in pages are not likely to be subjected for consideration for removal to the supplemental. Sites that do grow in size become subject to consiseration for removal to the supplemental index with those sites adding the least number of pages being prererred to remain in the main index.

Thus, by just tripling this site's size, it just became a likely candidate for being sandboxed.

The reason why I say that the main index is a set of whole sites is that for the purpose of matrix calculations it is much more efficient and the defining points are easier to establish if whole domains are included in a single index.

caveman

7:10 am on Oct 28, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I dunno.

I can't yet find another explanation for this, but I'm just having trouble believing that adding pages like this really hurts sites badly. There are millions of reason why a site might grow in 10% or 30% increments, rather than arithmatcially and evenly.

I'm hoping some more members will jump in and tell us this is a non-starter...

Maybe it's just a severely tightened dup filter.

Powdork

7:11 am on Oct 28, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Interesting northweb, how long ago did you add these pages? Was it all at once? Are they using the same template as the existing sections of your site? Have they been indexed yet? Any other info that might be pertintent as to why you have had no ill effects, yet caveman has?

Spine

7:34 am on Oct 28, 2004 (gmt 0)

10+ Year Member



"Maybe it's just a severely tightened dup filter."

Is 'tightened' the same as 'retarded'?

Iguana

8:11 am on Oct 28, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I finally gave up on trying to get a new site out of the sandbox and put a number of the pages on my main site (about 12,000 pages to a 3,000 page site). This was only 4 weeks ago and ranking/traffic for the main pages have stayed the same. I don't think the new pages are getting any traffic yet but I wouldn't expect much.

More likely is something to do with the glitch/readjustments on aug 5, aug 25, sept 25 - where sites plunge down the rankings for just one month

lufc1955

10:24 am on Oct 28, 2004 (gmt 0)

10+ Year Member



Being dumped for adding new pages seems so unfair. In our case we added nearly 50% more pages of unique quality content. It was the sort of content not found anywhere in the UK on any other web site. The role of a search engine is to make it possible for searchers to find this sort of information. But we were dumped out of the index and the searching public are worse off as a result because they can't find a site that has been around for many years and has unique news articles and quality content that has been built up over many years. The theory about the index being full is an interesting one and would explain why some big sites have seen the number of pages in G diminish.

jino

10:51 am on Oct 28, 2004 (gmt 0)

10+ Year Member



I have a site with about 70,000 unique pages for years now. Recently we added about 60,000 new pages over 4 months and have just grown from strength to strength. I honestly find it hard to believe you would take a hit for adding new pages unless those pages are linking to something dubious..

pardo

12:31 pm on Oct 28, 2004 (gmt 0)

10+ Year Member



might the capacity of the database be an issue, as mentioned in various threads in relation to the lag/sandbox effect lately...?

some say yes, some say not but there wasn't proof this IS NOT an issue.

also strange GG dont answer many questions as often as before.

saying nothing can 'tell' a lot also ;)

ownerrim

2:59 pm on Oct 28, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I'm not disputing this phenomenon might be happening but it does sound insane to think that adding content to a site can get it slammed overall. What the heck would be the logic in that? Sandboxing new pages if a large number is added at once doesn't sound fair either, but I can understand that better than an entire site getting thrown down a few hundred rungs on the serps ladder.

Someone else said this and I will echo it. Whether google is the cause or not, it breeds fear. Too many links, not enough links, link placement, anchor text wording, and now....the fear of adding content? Insane.

randle

3:06 pm on Oct 28, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Cat and mouse, on and on it goes. Just think what SEO was like 18 months ago. You could build a site, get good PR and be right where you wanted to be in under two months.

Now you think twice about adding new pages to sites that are 4 years old!

WebFusion

3:08 pm on Oct 28, 2004 (gmt 0)

10+ Year Member



also strange GG dont answer many questions as often as before

Not strange at all, since google is now a publicly traded company.

How would you like to be the one blamed for a 10% drop in stock price for a comment you made on a public forum? Thta might not soudn like a big deal, but to people (including many google employees) who own millions of share, that is big bucks.

No...now that the google (most) of the google employees have a vested interest in the highest stock price possible, I doubt we'll be hearing much "inside" information anymore (not that we got much in the first place).

Google simply has too much interest in keeping as much as possible about it's systems secrets, becuase:

1. The want a high stock price
2. Their "enemies" are gunning for them
3. The spam problem has grown out of their control, but to admit it would be harmful to their interests (see #1)
4. [theory] : Their database has reached it's limit, and they are frantically trying to develop one that can scale further (when was the last time the "Searching 4,285,199,774 web pages" was changed?)

This 84 message thread spans 3 pages: 84