Forum Moderators: Robert Charlton & goodroi
Over the past couple of days, the number stood at 999 using the site:www.foo.com command. Today I noticed something interesting. The same command that returned 999 pages over the last several days now shows 9,140 pages. Quite a jump. Now I understand why everyone's been talking about inflated page numbers. I was wondering if this happened only for sites in excess of 1,000 pages in Google's index. For someone that just crossed this milestone, that was my experience.
Hardly surprising, since one of Google's aims is to present relevant information *quickly*.
DerekH
Are there people out there with 1126 page websites who display 1126 pages via the site: command?
Having those inflated page numbers obviously has no positive, so staying below 1000 pages appears to be something to definitely aim at, if only to track how many Supplementals and other screwed up listings Google has for you.
The telltale sign for me is my name (uncommon last name). After coming out of the sandbox in June / July, I used to rank #1 for my own name (actually held the #1 and #2 spots. Now I've got about ten sites above me.
The site grows at roughly 5 pages a week for the last 7 months, but I'm wondering if the 1,000 page threshold is another flag for Google. Certainly if Google "thinks" the site is 9,100 pages, then it might be suspicious of the site growing by 8,000 pages overnight.
I have had a rewrite for the non www pages for over 6 months, but google recently brought back in cached versions of non-www pages from last year.
Seems like everytime they start indexing the site correctly, they bring in the non-www pages or other very old versions of the pages with various url. G index is showing some urls for pages I have never had.
Prior to doing the 301, Google showed different counts for www and non-www. The current count is pretty close on my site.
My site was showing 8,232 recently (a reasonably accurate count)
But now it shows 56,998 which is unreasonably high, I never produced more than 9000 pages for sure.
My link count has dropped to record lows though, basically my links disappeared.
This inflations in page numbers, is it happening ONLY for the penalized sites/ sites lost ranking on the serps, or its happening also to sites that rank well on the serps?
I have a site that is inflated from 1,200 actual to over 10,000. Not penalised and never (touch wood) affected by updates.
As with others, all my sub 1000 page sites have accurate page counts.
This started quite a while ago and the page count has risen 2 or 3 times since
Thanks for feedback.
>>I have a site that is inflated from 1,200 actual to over 10,000. Not penalised and never (touch wood) affected by updates.<<
So that "inflation" is part of business as usual for Google now!
>>This started quite a while ago and the page count has risen 2 or 3 times since<<
And that enables Google to tell Yahoo; Mine is 3 times bigger than yours ;-)
It seems strange, the reasons appear obvious but the ruse is so pathetically crude it is bound to get out into the broader public domain and will, presumably, be quite embarrassing for them
It's a bit confusing to me, but here is my logic...
Google knows the actual page count of any site in its index - it's just a calculated value (actually just a count). But why doesn't it show the real value? Is it too lazy to calculate the value and uses an estimate instead?
But how else is this estimate used? Clearly if it stored the real value, then it would show that in the site:www.foo.com query. So perhaps when a site gets large enough it uses an algorithm instead of actual count - perhaps to save CPU cycles. But if this estimate is now used as the "actual" for a site, then perhaps a penalty can kick in for getting too big too fast. Just a thought…
If I'm wrong, then of course Google should explain why it used an "inflated" page count when it could just use an actual value it already has stored somewhere. Personally, I'd rather believe the first explanation (I'm not big on conspiracy theories).
This inflations in page numbers, is it happening ONLY for the penalized sites/ sites lost ranking on the serps, or its happening also to sites that rank well on the serps?
I have a personal hobby site, that, as of the Bourbon update, ranks extremely well in the serps - pretty much #1 for all the words/phrases I want it to rank for. The site has ONE page; it's database driven, and the search results actually show on the index page. site:www.mysite.com returns 618 pages - so it looks like Google is seeing every database record as a separate page - except that there are only 390 records in my database, so I'm not sure where it's getting the extra pages. It's not a question of www vs. non-www - the non-www results show a different (lesser) number of pages altogether.
Weird.
It looks like that DC does not have my non-www pages index, but shows tracking urls and a couple of other old dead pages, still a lot more accurate than I have seen lately, though almost every page is supplimental.
Maybe they are moving in the right direction...
Under 1k pages, very accurate"
I see the same thing with our sites. On this main site (over 1000 pages - 5 years old) the directory structure has changed at a few points in time, url's have changed (were 301's and now 404'd or 410'd, both www/non-www pages were indexed at one time (although the fix has been in place for a couple of years), Google creating extra pages from 301 external redirects (since been removed but pages remain), screwed up links to our site... We show an inflation of about 8x
The newer sites have not had any url changes, screwed up links, redirects, removed page, www/non-www fix been in place from the beginning... These sites are under 1000 pages. Zero inflation for these sites.
I 301ed enough pages to get below 1000, and magically Google now correctly lists the pages on that site almost exactly right (a few non-existent Supplementals included) rather than the 10x number when I was over 1000.
No ranking improvement though, which may be the scary part.
I know at least one cause that will trigger this bug, without any doubt at all, but it cracks me up to watch google mess up this much so I'm going to keep it to myself for now, but I've verified it on a wide range of sites, and it's always triggered if this one condition is present. Other conditions may also trigger the bug, but his one definitely does.
Now I looked at the inflation and have more than 38,000 pages in the index. That is just not possible since it is a blog with less than 2000 entries. Even including comments and other stuff it wouldn't be 20 pages for each entry.
And I don't get a better result on that other server you mentioned.