Welcome to WebmasterWorld Guest from 184.108.40.206
Forum Moderators: mack
Until I stopped looking for my pages and really started looking at the results I was getting for my keywords...NOT EVEN CLOSE. Little to No relevance.
Looking for blue widgets in North Shores Florida...getting used car lots in North Chicago, or pet grooming in North Carolina.
So, if you've lost pages or placement...relax (it's probably not your site) and wait for the bugs to get worked out.
Surely MSN is working on this...
If you are experiencing the same or have comments, please post.
I'd like to know it's not just me... :-) Fish
twebdonny: No, the penalty doesn't keep the bot away; it will be back on its usual schedule, eager to see what you've come up with since its last visit.
As far as less is more, there are two sites ranking above me on MSN for my main keyword phrase that have very little content. One basically says, "If you are looking for "keyword" please click this link(to another website)" and the other says "This site has been disabled." Is this an example of less is more?
It must be very tough for SEs to try and deliver quality search results with millions of spammers constantly looking for any loophole in the algo to exploit. But it is also very tough as a website owner to try and build the best site you can for the user while seeing SEs rank garbage sites in front of you constantly. I do build my site for the user. But I can promise you that I know my user much better than MSN's algo does. And while I am thankful, MSNDude, for you posting here. I'm not sure how to respond to you basically saying, if you dropped in MSN's SERPS, stop spamming or stop looking like a spammer.
The whole idea assumes that MSN has it right, knows what the user wants and if you dropped you are doing something wrong. Are you really that confident in the quality of MSN's results?
For My Site.
Meta Tag Descripion...Repeat Main Keywords no more than 2 times. Secondary Keywords 1 time.
Lowered body individual keyword density to no more than 3%
for main Keywords and 2% & 1% for secondary keywords.
Overall page density 1%
I chose a conservative approach with 1% overall page density...you may want to experiment. MSNDude is right...
optimize for your visitors and you can't go wrong.
Spot on with your findings, i have done similar tests with similar results on some test sites.
However, your problem comes when you have a large page that may contain a lot of information about the different types of widget with perhaps a lot of links to other site pages that are about related widgets. In this situation your keyword density for "widgets" will rise significantly higher and you cant do a lot about it, i suppose you could fill the page out with more guff to bring the density down but you are then not delivering effective content for your users.
What this means is that to rank high in msn they are in effect encouraging webmasters to build low density pages of little content so that they rank - hence why the junk and spam sites clean up.
The junk wins because it has little real content behind it anyway, the spam because the spammers have nothing better to do than tweak density levels to get the desired results.
Meanwhile the authority content sites suffer because certainly on the sites we are involved with we are not likely to mess with reducing the density to please msn at the risk of losing position in google whilst reducing the effectiveness to the end users who like the pages marked out the way they are.
This is why i believe msns current algo will be scrapped, its just a matter of time.
So far, most of the examples I've seen of pages that got hurt by recent algorithm changes were pages that could made a few user-friendly changes and escape the penalty. If that's the case, then that's good for you, good for us, and good for our customers.
A big part of why I post here is that I really, really want to see counterexamples where our algorithm is getting it wrong. That's how we improve it over time. Don't hesitate to tell me that you think we've got it wrong, and don't hesitate to send me a "sticky" with a concrete example.
What I am finding is that a more natural language patterns rather than a keyword stuffed language seems to be doing much better. Sometimes content is rather weak sometimes it isn't. Alot depends on how good the content is for the subject is in the first place for the specific key phrase.
Also many webmaster have a habit of just focusing on their main keywords words in their page. Think about a natural language. If you talk about widgets as an authority you tend to use combination of words which gives the page context. Not talking about stemming here. Many time webmasters will remove or avoid such words to stuff in another keyword or phrase.
Let's say you write about pond widgets. Simply using the keyphrase pond widgets over and over does not add any real meaning to the page. What adds meaning to the page is all of the other words and language patterns you use to express the subject of pond widgets. That set of words are unique to the subject. If you were to replace all occurances of "pond" and "widgets" with the word "farfulgoobersteins" and "opootogglegooprewart" you should still have enough context to know what you are talking about "pond widgets". Try it and read your content you may see a HUGE difference in how the content reads. You may get annoyed by all of the mentions of a meaningless word as "farfulgoobersteins" and "opootogglegooprewart" you may not really have a good sense of what you are talking about through bad context. Again try it...
If you enter a site and it is about "pond widgets" and in the navigation it mentions "pond X widget" "pond Y widget" - "pond Z widgets" you can see where you can get excessive and unnecessary mentions of the word pond and widgets. Simply using X, Y, Z should carry enough meaning along with other content to know what the page/site is about.
I'm out of msn, I don't show up for any searches except exact url and exact title in quotes.
When I do a site: command I notice that the first result is a url that has never existed on my site, it's just https then my domain name. But I have never used a secure server.
Is the order of urls returned by the site: command relevent to anything or ...?
Also, a few years ago I switched from html to shtml and I put a 301 redirect from the old index.html to the new sitemap which is quite large. I noticed that with a site: command msn still has the url with the index.html (not the new sitemap)... so could that be part of the problem? Does msn think I have a humongous index page? maybe...
msndude I stickied you my url a few days ago I'll sticky it again in case you are interested. It would be good to have a fresh eye on what could be more user friendly - thanks :)
It sure does look like pages with content have been seriously demoted, and it also looks like the knobs have been turned down so far on number of occurences and exact match that it's pulling irrelevant stuff up that has nothing to do with the search term except for random mention of a word used in the search phrase.
I'm going to send to you another example of where I don't think your search engine should be able to return great results, but it's missing on a couple of them.
For this particular query:
1. On target, but not what I would consider a high quality result.
2. Another subpar result, but also the one that Google also likes for some reason.
3. From the same website as #2 and I really don't know why. Also favored by Google, so there must be some off pag factors fooling the engines.
4. Reputable company, but this seems like the wrong page.
5. An article that I would consider "decent." Perhaps the best result so far.
6. A result showing a Google.ru result's query. Very strange.
7. Right "industry" bad result.
8. Right industry, bad result.
9. A page redirecting to another page (not good results). MFA type page.
10. Not an "on target" result.
I picked this particular query because I rank well on Yahoo and Google for this term and it's because I've written a 800 word article on the topic - it's actually pretty good. However I rank around 29 in MSN for this query.
Regarding the "less is more" point - I find it hard to believe what im reading. If a site is an authority on a subject or it is a site in a large topic area, it is going to have a lot of content.
MSNdude, I have to agree with RichTC on this. When content is having to be cut in half just to maintain steady ranking, I think something is off. Your goal should be to index the world's content. If I have to cut 60% of the content on a page because the file is "too big" to compete with pages consisting of a few sentences, then something is lost, isn't it?
Instead of "sometimes less is more" perhaps it would be clearer to say "remember that spam is often about excess." If you are doing something to excess -- something that you're hoping will impress the search engine, even though you know it'll annoy customers -- then you should reconsider. That's all I'm saying here.
By the way, I'd also like to point out that about 80% of the sticky notes I've received on this topic have come from people whose sites were completely unaffected by the change.