Shaddows - 7:54 am on May 4, 2010 (gmt 0)
Here's the Geek Question: if you were a Google engineer and suddenly had all this extra data storage capacity, what would be at the top of your list to implement into the Algo-stuff you had wanted to do previously but couldn't because of the data limits?
How about, making personalisation work? Feed everyone A/B testing, find demographic prefs, exploit.
FWIW, the concept of reducing the megasite Gravitas Constant* roughly correlates with my data (at least, its not inconsistant). Pages that "just ranked" for mid-tail are not stable.
Can I encourage some other data-crunchers to start looking in the wilder regions of volume keywords (past page 5)? Churn has increased. I have a sneaky suspicion that two things have been devalued- internal linking and "TrustRank". It's rippling through midtail and lower-order volume SERPs, and I reckon it will hit the big time soon.
Incidentally, Vince et al could have been a precursor to this, in protecting Branded (or otherwise "expected") results from the ensuing disruption.
*Like the Gravatational Constant, additional mass (pages) automatically increases attractiveness