Forum Moderators: Robert Charlton & goodroi
The update now appears to be over. So let's discuss the algo.
What I think google changed in the algo this update...
- Backlinks count for less. I see pages with a handful of less than outstanding quality doing well.
- Keyword density counts for more. I see poorly linked pages stuffed with a set of keywords doing well.
- Semantics / LSI - no evidence. If anything I notice a narrowed focus works better. 'clothes for women'!= 'clothes for her'!= 'clothes for woman' in Google right now.
- Hilltop = no evidence. Seeing pages from small, obscure, never seen before sites ranking well.
- On page factors count for more. It's the only way I can explain some pages rankings.
- No evidence of penalising affiliate sites.
There's a start... my tuppence worth.
I look forward to reading your analysis / thoughts.
ps: mods, please help keep this thread clean by deleting rants / moving to other Allegra thread.
I was in the sandbox for the last 6 months, and now it looks like I may be out of it. Since some of the data centers show that I appear to be out of the sandbox, does that mean I probably am out?
Ran a little test (we have a 40,000 page site PR7 index page and very very hard hit).
Took one inside page (but linked from index page), one commercial search phrase (usually 750000 results only) for which we ranked top 5 before Allegra and >70 after Allegra and nearly tripled the keyword density on it. Nothing but nothing else changed.
After last spider (so with a fresh date 9th Feb for sure) now ranks 3 on g.com and g.co.uk. Other pages in SERPS look identical'ish.
Seems to simple to be true - we rank third on our company name - but our index page doesn't actually use our company name much. Fix that tonight and post back tomorrow.
Also
<half-baked theory>This is really all about "local". The difference between post Allegra ranks on G.co.uk and G.com is much wider than pre Allegra. - ours is a UK site.</half-baked theory>
If you have 8 Billions of pages and some have just not ranked before - how to bring them into good positions without completely killing others, well ranking pages? Think about it, what would you do?
If you have thousands of people optimizing their pages and building links to increase their rankings - and you don't want that. You don't want all those realiable kw analyzers which are tracking thousands of keywords.
What would you do? Imagine, youre Google and you have to please Millions of Webmasters - and pleasing is keeping your importance because John Sixpack wants to see some referals from Google. Imagine you have an index that has grown very fast. And imagine you have millions of users daily.
So. What you do? Know what i would do? Building several different indexes and randomize the search results to split your traffic. Still serving relevant stuff - but there are always thousands of relevant results which you may serve to users looking for informations. I do believe Google is doing exactly that - and that we might throw all keyword tracking tools away :-)
itloc
[edited by: Brett_Tabke at 3:51 pm (utc) on Feb. 10, 2005]
[edit reason] off topic [/edit]
It has been said many times (and I agree) that if the serps aren't stable it is bad for searchers as they tend find a site the second time by remembering the keyword they used as opposed to adding the site to their favourites. This is just very confusing to the average joe surfer. I think things will stabilise within a week or two.
If it isn't related to this my other theory is it is related to IP addresses and cross-linking as both sites are on different servers.
Google seems to be penalizing sites on searches for their own company name possibly due to overuse of the name in inbound link text (lol). If this penalty has a general component (if it is not entirely keyword specific) then we might expect to see fewer home pages in this index since these pages would be the most likely recipient of a company name overuse penalty.
hahahaha. Google penalizing us for bad SEO. Sorry Google, will try to stuff as many keywords as possible in the anchor next time.
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0 Transitional//EN"
"http://www.w3.org/TR/REC-html40/loose.dtd">
I just compared DCs with the McDar tool and I saw my sites ranking very well on nearly all DCs. But when I try to search on google.com, my site is not there and the results are different.
Taht means Google shows different results depending on if you look in the frame of the mcdar tool or on the real datacenter in an own browser window. But why?
After some research I found out that the mcdar tool uses the following format: [216.239.59.104...]
But if you search regularly on the DC the search URL is [216.239.59.104...]
The difference are the twowords "ie" and "search".
When searching the "ie" style I rank very well on nearly all DCs, when searching the "search" style I do not rank anywhere.
What'S this all about?
greg
The &hl=de variable means that the search is German language. The McDar tool states no language so &hl=en (English) is assumed.
Err, the /ie means "International English" and you get taken to a site that says "Google English" rather than the usual "Google" as seen at www.google.com.
yesterday allinurl:mysite.com produced 60,000 results and only 350 or so were not in the supplemental results
also yesterday allinurl: and site: searches were returning different results.
now allinurl: and site: results return the same number of results, now only 5500 or so and I have nothing listed as being in the supplemental results
Err, the /ie means "International English" and you get taken to a site that says "Google English" rather than the usual "Google" as seen at www.google.com.
Not true.
[216.239.59.104...]
"They show title-only, but the snippet is there via a mouse-over. This interface is the /ie interface."
The ie interface has nothing to do with the IP addresses.
[216.239.59.104...]
[google.com...]
Whatever, when I visit a direct IP address, I see that directly underneath the "e" of the big Google logo there is the word "English". I don't ever see that at www.google.com.
"All Google IP addresses have a stripped-down interface with no ads."
www.google.com is nothing but a domain name that gets resolved to one of over 50 IP addresses. The Internet uses IP addresses, not domain names. Yahoo is different -- they intercept and redirect you unless the domain name is in the packet, even though you have reached their site with just their IP address. They've been doing this ever since the worm problems of last July.
All Google scrapers use the /ie interface because it is stable (the html code has never changed) and low-bandwidth (no ads or extra coding).
It doesn't always take you to a doorway page with "English" on the top. There are over 50 Google IP addresses you can try with the /ie interface, listed in other threads on WebmasterWorld. Try them all yourself and see.
>> There is no such thing anymore as a "dc". There are just some ip's which are serving different results - just like google.com. <<
[webmasterworld.com...]
thanks for the clarification. But in that case something is really wrong with Google. Why are my sites showing on prominent position when searching for english sites when they are obviously german. The language is german and I added a meta-language tag.
When searching in german google the sites are nowhere.
But for example on [64.233.171.99...] I am found on the good positions without changing the language.
Really frustrating. Made some good content and good optimaziation but Google gets my page in the wrong language.
greg
20 primary keyphrases - mainly 'medical', not fiercely commercial.
Top 5 positions for these phrases over nearly 6 months, variety of pages.
Good keyphrase density, keyphrase in title, good anchor text, all pages 'static' - no dynamic elements to them
All 'white hat' - no dodgy techniques.
Now, all completely gone! No much coming up on a 'company name'/domain search.
My guess is this is either a totally revamped algo, or as I have seen previously, a 'work in progress' database.
What to do? Wait, hope and trust that Google will put this right.
I am seeing so many SERPS containing news releases from portal site, etc. that this cannot be providing good quality results.