Forum Moderators: open
Modifications of the PR formular (as done by Google in the past) cannot explain the results. Also, even an incomplete update wouldn’t explain all data. Thus a change in the iteration scheme seems to be the most likely explanation.
This behaviour might explain other phenomenons such as the 'Google lag'. Of course, this is pure speculation, but the inconsistencies of the PR calculation (sprictly speaken, the inconsistencies of the PR values shown in the toolbar) are fact.
DJGreg:
> Why would Google link to directories rather than to the content sites.
Because this new algo might be a first (?) try to cope with the "authorities"-stuff. Note that the dmoz seems to have completely vanished from the backlink lists. Seems to me the authorities will be hidden from the search results in the future, and the algo is far from perfect in deciding whether a linklist is an authority or spam. I think many of these lists got PR0 now and will be thrown out completely the next update.
> So why would these sitemaps, which only consist of links rank better than the content sites?
Because they still - according to the old thumb-rule -inherit a PR-1 form your index-page, "send" a PR-2 to your content pages and the external inbound links of the latter do not suffice to switch this up to or beyond PR-1 again.
> "topic sensative" anyone?
I see no post off topic:
>>> Google’s PR algorithm and/or iteration scheme changed … and it seems that they have a problem
Faced with obvious massive changes this is a naturally chaotic period of collecting empirical data and formulating first theorems. The goal will be to either find more precise coefficients to mirror the behaviour of the PR-calculation-algorithm or to find new fields of activity for all then unemployed SEO-experts *g*.
If Google have changed the algo, that could explain the lack of PR update over recent months. Perhaps we'll see regular updates again and inconsistencies will gradually diminish.
What you say makes sense, kaled. There already is enough complication and obfuscation in the serps algo. There is no need for Google to obfuscate or complicate PR calculation as it's just one factor in many that affect the serps. PR as such is a pretty straight-forward concept. What they could have changed are
1) the Coefficients (dampening factor, logarithmic base of toolbar PR) or
2) the Computational algorithm. The PR formula for one page is a linear equation. The PR formulas of all pages combined is a giant system of linear equation. It takes quite a while to solve this system. I'm not good at math but what if they changed the PR calculation to something inkremental that allows for almost instantaeneous PR update?
It also seems though that some of the discrepencies are so completely unearthly that the *most* logical explaination would be "screw up".
I see zero evidence of topical sensitivity or any consistently different dampening within a domain. Pages that are exactly as relevant as peer pages with the exact same linking sometimes show inexplicable differences. A mathematical equation does not allow some of these things to exist. Maybe a bollocksed search engine does though...
<but just to reiterate again, I only have a dozen examples, but all could be explained by an age-of-link reason. This is the single they they have in common.>
[edited by: steveb at 10:38 am (utc) on Oct. 8, 2004]
Interesting thought. In view of the patent issued not too long ago, some are wondering whether there might be intermediate iterations based on Local Rank, which could narrow things down topically. Even as far back as Florida, there were sites that were seriously filtered, not having any inbounds from topically related sites or pages.
My own experience has conviced me that on topic backlinks confer a serious benefit now.
This is a change from the past but I wouldnt call it radical. Actually, I think its fairly evolutionary.
Back links purchased from off topic but high PR sites become worth a lot less. A great way to cool the link market.
It is mathematically impossible for www.site.com/ to have a PR3, even if it is possible for site.com/ to have its own PR6. (The genuine situation should be www.site.com has PR7, while site.com has a PR6.)
This is a textbook new/lagged domain that got tons of both high quality and low quality, and both high PR and low PR, links to it.
It would seem hard to not conclude that the main page, but not inner pages, of this domain got a PR devaluation because either the links were new or (very unlikely) that it somehow got way too many links from way too many sources.
My own experience has conviced me that on topic backlinks confer a serious benefit now.
Yes, but what topic? Google can't guess the topic of a page out of the blue, can it? PR is a simple equation that takes links into account and nothing else. It would be easy to filter the links that go into the equation by say time in order to have PR indicate sandbox. But it is highly unlikely that something blury like topic is considered for PR.
The topic comes into play is when the user specifies it by entering a search query. Then, on-topic links count in the form of LocalRank and anchor text. But again, on-topic PR is out of the question.
This would seem to make sense as the number of sites is growing at a phenomenal rate, and (probably) so too is their quality. Making it more difficult to climb the PR ladder might be Google's answer to this problem.
PR does have a real effect in so much it appears to influence the amount a site is crawled. For example, by pushing down lower end PR5 sites into PR4 would be a significant saving to Google.
Yes, but what topic? Google can't guess the topic of a page out of the blue, can it?
If you have an adsense account try pointing it at one of your pages and see if Google can figure out what the page is about. If not try one of the adsense preview tools. It may take a day or two but it will get there.
I'd say statisticaly, for all practical purposes, google is very safe in assuming that each page has one topic. This doesn't mean that each page only has one topic, but on average it does, high above average would be my guess for ranking sites. I just checked using the brett WebmasterWorld theme tool, it got a site I often compare things to themed correctly.
It could also just be that the highest PR page out there has increased absolute PR more than yours. With a true log scale, in this case your absolute PR could increase, yet toolbar PR decrease.
If Google have changed the algo, that could explain the lack of PR update over recent months. Perhaps we'll see regular updates again and inconsistencies will gradually diminish.
Hopefully you’re right.
It is mathematically impossible for www.site.com/ to have a PR3, even if it is possible for site.com/ to have its own PR6.
That the kind of 'inconsistencies' I'm talking about. These things cannot be explained with a change in the logarithmic scale, a change of the damping factor, a devaluation of some kind of links or something similar.
Also modifications of the PR formula are not so easy as it looks like. You have to ensure that you’re ending with a mathematical valid model.
speculation: 1. …8.
As explained above, normal modifications couldn't explain the data. Even a random factor (multiplicative or additive) for PR value shown in the toolbar wouldn’t be consistent with all my measurements.
A random PR source for PR calculation might be a solution.
Are you using the PR display in the directory or the toolbar? I would venture a guess that the toolbar PR has nothing to do whatsoever with a pages real PR and is showing some display similar to or linked to the new backlinks display,
I'm using the toolbar PR value. In the past these values were related to the real PR value in a logarithmic way.
The concept topic is something way too fuzzy to be considered for something as plain mathematical as PR.
Lets say your page is about fuzzy pink sock puppets. Using the same technology used for adwords keyword suggestion tools it's not that hard to come up with a list of related keywords for the topic of your page. So links out to other fuzzy pink sock puppets pages could get the highest percentage of PR transfered. Links to puppets, socks , or pink could get a slightly lower percentage transfer. Links to blue widgets would only get a small benefit.
DJGreg:
> Why would Google link to directories rather than to the content sites.
Yep seen this in abundance in my SERPs, directories are all over the shop. My theory is that most directories have sponsored links provided by PPC engines (including Google's adwords). It would be unethical if Google only boosted the dominance of directories that carried Adwords so they just boosted the prominence of all. Great for the exposure and promotion of PPC culture. Google won't plaster their own SERP results with Adwords but they might get away with it if the top listings in the SERP’s were from directories containing sponsored results.
Personally I was hoping to see a reduction of directories cluttering up the SERP’s
I think you throw up a good possibility.
"8. Throwing in a random obfuscation factor in Toolbar PR that randomly affects some pages and not others (to confuse SEOs)"
First they confused the heck out of us by displaying disinformation with the backlinks, and now maybe they have instituted a program of disinformation with the tool bar PR display. If the disinformation factor is different for different people, makes collective efforts like this board Googles best friend. Everyone comparing data, observations, anecdotes trying to collectively figure this out and its all just bogus. Everyone is having a different experience.
The back link display, and the tool bar PR display, (however late and inaccurate) definitely helped me inch my way up the rankings. Maybe there both gone now, it won’t stop us, but they were excellent aides. Makes sense to me they would take those two bullets away. Sounds a little goofy I admit, but it, "just keeps gettin harder every dayyy"!
This way, it keeps out people who purchase front page links for X amount for 2 weeks, for example.
The links that have been in place for say 6 months, give a higher PR than a link that has just recently been spotted by Google.
I also agree with on theme PR counting for more, for example, a Jesus Christ followers website linking to a Viagra shopping site.
Steve
Even a random factor (multiplicative or additive) for PR value shown in the toolbar wouldn't be consistent with all my measurements.
For pages on my site which are undervalued, those which are linked in similar manners share the same pr.
In one example
Article X shows pr 6
~10 Article X_Y (with X as the parent and only linking page) show pr 1
~50 Articles X_Y_Z (with one of the Y's as the parent and only linking page) show pr 4
One Y is older and shows higher pr (5), and incedentally the Z's are newer then any of the Y's. The anomolie Y's were created ~4 months ago, prior to the 2nd to last update but were new enough by a few days not to show pr at that time.
2) OT: [news.google.com...] is this new? Never seen that before and a lot of beta-pages on it.
> My own experience has conviced me that on topic backlinks confer a serious benefit now.
I'd support that. Recently I wrote a little script to analyse my backlinks as compared to my competitor's on a given keyword. I thought those directories which link to most of the result urls of a given search-query would make up a good approach towards "authorities" and spend pretty much time to add my url to those directories, that had not listed my site yet. First there is a massive change in the results of my script between sep 18th and the one run i did today. Secondly the result of all my efforts was quasi null, whereas a good friend of mine, running a site on plants and preferring qualified, content-related backlinks has made good progress the last weeks.
As to the technical arguments: I see no need the PR algo itself had to be changed. Its no problem to let a number of additional (semantic) filters run over the first PR-table and present the result of these operations to the toolbar and serps.
<but just to reiterate again, I only have a dozen examples, but all could be explained by an age-of-link reason. This is the single they they have in common.>
My examples show just the opposite results that would be attributed to an age-of-link benefit or penalty. It looks more like a have-baked calculation. They didnt complete enough iterations, or the calculations left out data. I dont see how anything other than "fuzzy' math accounts for steveb's presidential site example.
My website promotes a company name .. Google doesnt know by looking at the company name what my website is about
You are selling some sort of product or service I assume, and hopefully you are going to talk about it. Not saying you can't fool it, but why would you want to?
Try the adsense preview tool
[webmasterworld.com...]