Forum Moderators: open
One of the key Hilltop concepts was identifying "affilated" websites - sites whose IP addresses shared the same c-block, or sites with the same right-most token in the domain name when the TLD is stripped (orgname.org and orgname.com would be considered affiliated and links between them devalued.)
Has anyone seen actual evidence that one or both of these "affiliated" determinations are actually in effect?
Which brings us back to your original question.... does anyone have any proof?
Which brings us back to your original question.... does anyone have any proof?
Unlikely. We can only use our analytical skills to deduce a probability. This discussion is very useful for that purpose.
Keep in mind that these papers were written a few years back and what was then a good notion for detecting affiliation may no longer be the case. The important and unchanging fact is that detection of affilation is an absolute requirement for these algorithms to work. That means a lot of attention would be given to that area.
If you think it through there are many on and off page clues to affiliation rather than just the server's C-block. My guess is that they would use a set of these clues and assign a threshold score at which affiliation is assumed. In the absence of other clues a shared C-block might not hit the threshold.
IMHO, these two topics are closely related. Having a number of sites that were never sandboxed this year, and others that were, we're able to assess the differences between the two sets. We see strong connections between these two topics.
But maybe it's all the houch we drink here in the cave. :-)
We consolidated 4 ethnic specific domains into one portal which got its dmoz listing fairly quickly. The 301s were from sites which have existed since '97.
since hilltop kicks in for broad topics with a lot of pages, i dont see any negative aspect from google's point of view by disregarding "suspicous" backlinks and loosing a few legit pages. there are still enough pages left.
When we look at our '04 launch sites that had no issues, and those that did, and then consider that most who post about the problem are launching their sites with help from ... umm ... closely related sites, it makes sense to me.
I get hives when I think of all the ways that G can make connections between sites (e.g., IP, C, WHOIS, toolbar, SiteA->SiteB->SiteC->, dup code/content, etc.). Then I think about how their logic is probably that in most cases those connections are *not* coincidental, and that they probably don't mind if a few innocents get taken down in the process, for the 'greater good' ...
Well you get the idea.
If HT-style link quality evaluation were crossed against age and number of links, with exceptions for sites that meet certain other criteria, IMHO, you'd have the current landscape.