Forum Moderators: open
While GG was nice enough to give us a heads up on what to expect with Dominic in terms of time frame, and what would happen when, I don't think we really got an idea of why this update is taking 2 months. And we can't expect to have GG just tell us G's secrets. But that doesn't mean we can't speculate.
Q1. PR Calculation?
A1. I don't think PR takes this long to calculate. Even if they have a completely different way of doing it. So that's out.
Q2. New algo's?
A2.Well why wouldn't they just test it on a test box and keep going like normal in the production environment?
Q3. New Spam filters?
A3. Same answer as 2.
Q4. New system of rolling updates?
A4. This was my assumption until today. But then I realized, the freshbot is acting a little more like a deepbot, true, but it hasn't really gone after new content and again, why couldn't they use a test environment for this?
Q5. Transitioning to new rolling updates, new PR calculation, new algos and new spam filters?
A5. Perhaps, but again, why not use a test environment?
Do you see a pattern to these answers? Correct. It all points to the question, Why isn't this whole update in a test environment?
I don't think G screwed up. It's intentional. If not, then GG wouldn't have warned us. I think, though it's purely speculation, that they have done testing in a test environment, they felt it worked well enough to go live. But they know the testing and tweaking could not be done completely and properly just with the test environment. They needed the resources of all those PCs to test with full data and full resources in order to get it right. They just couldn't go any further with their test resources. So we are essentially witnessing a live alpha/beta test for that reason.
In regards to your question as to why they decided to put out an old index/new algo, it may have something to do with agreements with their partners like Yahoo, who might expect a new index once in a while.
In water filtration we have our raw material which is the Muddy Water {mud can be compared to our good old SPAM}. The most basic way to filter water is the Filtration Process. It involves filtering the water by using cloth like filter to remove most of the mud from the water. Then the other processes like percolation and UV rays etc is applied. But then you realize that the muddy water can be filtered better if you add an extra filter during the filtration level. Remember filtration is the first level.
To implement that you add an extra filter coating and start with the raw material from the beginning and then proceed to other levels. That's what they are doing now. Applying filters on the first level and then after weeding out unwanted stuff build the index by bringing in other factors like backlinks etc
What was happening was they were applying filters on the processed data for QC. I think the filter they are applying now is on the initial process of buildig an index. And that's the reason they are building the index from scratch. They will observe the results of this filter for search results. There will be many Poor Quality results in the initial stages because not all backlinks are accounted for as yet. Remember GG asked for feedback for SPAM after the Data centres stabilized. That feedback would directly go to QC and they will devise new filters to see how to combat those sites.
This is what i have assumed by reading all the threads just before dominic started. Hope i am right :)
Q4:
the only real problem I can see, is the high number of error 404 pages, which should be delete from the index.
googlebot is very diligent in spidering new pages and updating the index.
Q1:
it is same question for me.
may be they actually test the rolling update of the pr and link matrix.