Forum Moderators: Robert Charlton & goodroi

Message Too Old, No Replies

How to tell if Google likes a web page or not?

         

dataguy

10:49 pm on Nov 26, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I've been wanting to start this discussion for a long time. Two years ago I was hit by a Google penalty and lost 90% of my Google traffic for 7 months. I tried everything, even flied across the country to speak with a Google employee to try to get things rectified. Nothing worked until I decided to try to read Google's collective mind and eliminate the pages that Google disliked on my site. Once this was accomplished, our rankings returned within days.

My main site is mostly user-created content. We employ a team of editors to try to eliminate the SPAM, duplicate content and the content which links to bad neighborhoods, but even with this intent, it has been impossible to really know exactly what content could get us penalized.

Because of this, I've implemented a set of functions which try to guess which pages Google really doesn't like. The formula basically relies on Google's list of sites that they don't like to display AdSense on, and a mechanism that determines how much Google traffic each page has received over a given amount of time. Basically, if a page doesn't receive 1 visitor a month from Google over a 6 month period of time, the page is removed.

We've also looked at Googlebot crawl dates. The vast majority of my pages are crawled every week, but some pages are crawled at much longer intervals. After 60 days, they get the boot.

I realize that this is a different way of looking at SEO, but it has worked very well for us for the past year and a half.

I would like to know what other ways there are of determining if Google likes a page or not. Any suggestions?

Quadrille

11:06 am on Nov 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



As you know, Google does not 'like' or dislike' any pages, and I'd suggest that it's not a useful way to proceed, because it puts the emphasis on google, rather than you and your pages.

If you make good clean original pages, and you have some quality incoming links, then your pages will mostly do OK, and with a bit of work can do very well.

And Google helps you with copious webmaster tools and guidelines, including Google's SEO Starter Guide [googlewebmastercentral.blogspot.com].

Having a working knowledge of the guidelines, while striving to build a quality site, is probably a better approach than trying to second-guess the algo - which is an almost certain invitation to sick site syndrome - overdoing, overobsessing SEO, and simultaneously making the site unwelcoming to human beings.

dataguy

2:42 pm on Nov 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I'm afraid you've missed the point of my post Quadrille, and Google most certainly does have pages they like and dislike.

Managing a large (50,000 pages +) user-created content site does not allow for finely-tuned control. I can have two pages containing articles on the same subject with one receiving an avalanche of Google referrals and the other nil. There are reasons why this occurs, and I would like to explore those reasons and how to detect them.

Using Google's Starter Guide is a given, but having 15,000 members, all with their own agendas, using best practices is not realistic. Managing this across 250,000 content-heavy pages, one page at a time, is a completely different challenge than managing an ecommerce site.

If a web site contains too many pages that the algo doesn't like, the other pages will suffer from lower general rankings or possibly a site-wide penalty.

I realize that this isn't how SEO is typically viewed, but for certain web sites, this kind of reasoning can mean the difference between ranking well (and earning a small fortune) and being forced out of business due to low rankings.

tedster

5:46 pm on Nov 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I appreciate that user generated content has its own specific challenges, and your move to automate your decisions is interesting - as is you experience of seeing rankings return after beginning to operate this way.

You have listed two signs that Google may consider certain pages non-essential. On a purely information or ecommerce site, I would probably not automatically remove such pages. I would first do a manual inspection of the situation, but I appreciate that the situation with user generated content can be different.

I have been thinking about your question, which I assume is aimed at finding other criteria that you could automate. I can't come up with any, and if you are currently enjoying good traffic from G, then I would probably not add anything to your current actions.

When you took the first step and you rankings returned, how many pages did you remove? Also, even though the timing for your return to ranking well is quite suggestive, I'm wondering. Do you think perhaps you located just a few problematic pages that made the difference for you - and the rest of the pages you removed were not truly part of the problem.

dataguy

5:59 pm on Nov 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



In this specific case I removed about 40% of the pages, which amounted to 40,000 articles which had not received any Google referrals for at least 8 months. These articles where mostly very old and outdated, and many of them linked to MLM programs or affiliate products. It could have been coincidence that the penalty was removed a few days later, but after 7 months of being penalized, the connection appeared obvious.

Over the past year and a half we've experimented with other ideas for weeding out the bad pages and in the process removed another 15,000 pages. During this period of time, our traffic from G has doubled.

Just to be clear, we're getting twice as much traffic (compared to pre-penalty times) with 55% less content. This is why this topic is important to me.

tedster

6:03 pm on Nov 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



many of them linked to MLM programs or affiliate products

There may well have been some "bad neighborhoods" in that group of outbound links, right? Outbound links tend not to just develop "link rot" these days and go 404, but the domains get repurposed, and often not in the best way.

londrum

7:16 pm on Nov 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



sounds like you're putting all your eggs in one basket. you don't know for sure whether your system is working yet -- the traffic coming back might have been for another reason entirely.
if google suddenly dumps you again you will have needlessly deleted a load of pages from all the other search engines.

it would make more sense to noindex the pages for googlebot only -- and let all the other ones carry on. you can do that with just a simple meta tag in the header.