Forum Moderators: Robert Charlton & goodroi
I think Google is weeding old or stagnant pages out of the index to make way for new pages, it is the only way they can keep up with the internet IMO. I recently did a search for a topic from 2002 and it was like going back into the stone ages in search. Everything now is what is happening today, not years ago. I don't know what all your sites are about but even on the top sites it seems they weed the pages.
For example I ran a search for an electronics product from 2000, only 8 years ago. You can barely find traces of it in the sites I searched via Google. Now do the same search from a product from today, say the iphone. There is probably a billion pages on that. Now I am not saying they are doing things wrong, but with the millions of pages added every day to the internet they have to delete or else run out of space perhaps. I just wish they had the ability to search the archives easily for the topics or products that are "old". Right now you can do that with Google news but not Google search.
Anyway my point is I think Google looks at a site and compares all the content, then keeps some of the most recent content in the results including the higher PR stuff and puts the older stuff in supplemental. That is only a guess but seems to be what is happening.
Since the older stuff I looked for was probably dropped into the deepest parts of these sites I couldn't find it with Google anymore.
Maybe though this is the way the internet search will be, you use if for todays content only. If they had to archive all our sites I don't think it is possible, not with all the pages being added.
[edited by: tedster at 4:46 pm (utc) on Sep. 1, 2008]
Keyword1 + Keyword2
An algo can weight the first more than the second, vice versa, or anything in between.
I remember back when Google used to focus on the first word putting much more weight on it. You had to score high on Keyword1 to get good ranking for the next word (Keyword2).
Current guess: if your site has many pages which include Keyword 1, yet not necessarily right next to Keyword 2 in those instances, Google "lifts" (slightly) your ranking level higher for the phrase Keyword1 + Keyword2.
In other words, the algo compares the single incidence exact phrase against the multi incidences of the components of the exact phrase. (Theme Referencing.)
Or a dial shift to increase ranking based on more concentrated established theme.
At the same time I may be seeing a dial down based on words that seem more spammy on a site.
Another one of those annoying fine-line Google algo issues?!
p/g
My website has not changed much recently, so I'm sure Google has tweaked somehow their ranking algorithm.
I don't know what it means but it might mean something to someone.
All the best
Col ;-)
I know what I'm talking about. Most of these site trip either the co-occurrance filter or the cookie cutter (thin URLs) filter.
So beware making huge sites until you have the content to support all those URLs without creating thin pages.
Dump 89,500 of your pages, redo the navigation, and hope for the best.
The duplicate content is very useful to the visitors. Because we are combing content from 10 different websites to generate a mash up where the content become very vital for someone looking for that content.
I know what I'm talking about. Most of these site trip either the co-occurrance filter or the cookie cutter (thin URLs) filter.
SEOPTI Thank you for the information.
SEOPTI - our website is not a thin URL, it does provide good value to the visitors. How do google differentiate between a good site and a cookie cutter website. If this is done by some manual editor in some 3rd world country, who has never seen America and whose qualification is as good as dollar an hour, than definitely he will remove us from the serp because he will never understand the value which our client is providing.
The client is a startup company, so they can't effort to just rewrite the 89,500 pages of content just for search engines.
Here is my advice for larger amounts of data
- many deep ibl's
- semi to fully unique value add content
- avoid templates sitewide
- sensible site map
- unique title tags
- unique meta descrips
- more deep ibl's
- huge root pr
without most of the above you will never support 100k urls in the primary index.
The client is a startup companyHere is the deal right here it takes time and with a starup of 100 k in pages with 95% duplicate no wonder your gone.
Serious you really need to do something with the dup content being a new site it won't ever get off the ground without some serious and I mean serios links. The amount of links it will take will more or less then flip another filter as your building to many links to fast.
Best advice I can see is keep the 95 k in pages in a different folder just keep it from getting spidered viva robots text and the 500 unique pages get them up there to draw links, submit to directories, and do whatever you can to get links for the site in a controlled manner.
The 95k in dup pages will be the death of the site unless ya do something about it.
- many deep ibl's
- semi to fully unique value add content
- avoid templates sitewide
- sensible site map
- unique title tags
- unique meta descrips
- more deep ibl's
- huge root pr
Thanks for your good information
Currently we have semi unique title tags. First part of the title is from our database and the next part is common for a particular category of pages. We are working on making as many pages with unique title as possible.
Does ibl means inbound linking?
The amount of links it will take will more or less then flip another filter as your building to many links to fast.
We are worried about the same thing. As a lot of money is invested in this project, client don't want to buy any link at all.
The website is 1 year old, we came out of sand box in six months.
We are getting good traffic from MSN and Yahoo. We have good repeat visitors too. But client is worried because our Google traffic diminished a lot.
Best advice I can see is keep the 95 k in pages in a different folder just keep it from getting spidered viva robots text
Client wants to get rank for those, he believes that the information of this is very important and with our mashup we are adding a lot of value to the content.
For example -
We had unique content on new york widget and we were ranking for "New York Widget" for that page. But now, we have no rank for that unique content widget page.
We still rank (or recently got rank) 180 for the keyword "New York Widget" for an inner page that talks about "Manhattan Widget". "Manhattan Widget" is a mashup page and is not all unique content, but it is value added content.
I see this patter for all the keywords. Unique content lost rank, inner pages with not all unique content (not very related pages) still have the rank, but the ranks are 100+
Anyone sees simillar Issues? SEOPTI - Can you check if you are in simillar situation?
I Thank you all for your help
Thanks
Mbwto, yes I lost ranking for almost all of my sites on June 4, I'm convinced it's a new cookie cutter and/or thin pages filter.
After adding NOINDEX to the thin parts some sites are coming back.
Of course you will get less traffic this way.
The duplicate content is very useful to the visitors.
The SE's might not see it the same way. They might see it as being scraped material that is posted on your domain only to scoop other's traffic (the domains it was scraped from).
Sorry, bredren, if you've only got 500 pages of unique content, that's all you should have online.
We have lost most of our top listing on Google on the 20th June 2008 for most of our keywords. We had top 1-10 ( page1) organic listing for many keywords on our web site that disappeared over night. Not even for obvious keywords such as our company name, domain name etc.
Our site is not black listed by google. There are about 1000 pages indexed by google. But we don't appear on the first search results page as before. It appears on the last page or 5 ,6 pages on google SERP.
We have no idea why ? So we have no way knowing what we are doing wrong and do not know how to correct the "problem" !
Is there anyone who knows ?
As we are aware we haven't done any thing to violate the google webmaster guidelines.
We'd be most gratefull if some one could advice us what we should do to improve our ranking.
Thank you
Kind Regards
Tim
[edited by: tedster at 10:57 am (utc) on Sep. 10, 2008]
[edit reason] moved from another location [/edit]
What you describe has characteristics of two problems. One we've been calling the "-950 penalty" and you can read about it here: [webmasterworld.com...]
As much trouble as the -950 penalty (over-optimization penalty) is, the second part of your report is a bit more disturbing - not ranking for your own domain name. That usually shows that your domain has lost Google's trust, and that often happens if they suspect you of buying or sellng links that pass PageRank. It can also happen if your site has been hacked and the hacker placed hidden links on some of your pages - that's known as "parasite hosting".
Keyword specific 1 page drop penalty after human review [webmasterworld.com]
Manual checks are something I've never been able to figure out so I can't say. I'm guessing it's a similar type of scrutiny algo as with a competitive keyword.
It's not unreasonable, really, for the same algo to be applied to highly trafficked pages even if there isn't competition based on the principle that Google wants to satisfy users. If the content was poor, they would be sending many users to junk pages... But they don't have too much to judge a site by relative to similar competition, so they would have to use other signals.
On a different keyword note, I just found my old 950d site get a rebound for a top single kw to almost top 10. Meanwhile for the main kw in the industry I haven't really seen any recovery yet or minimal (ten places better at best), leaving it near 90.
Not that I'm too worried, though. For this particular kw, if you're not in the top 5, you're toast. The first five sites give most users the data they want, so they don't bother to visit the others.
I just noticed in my sector Google revised its related search suggestions to make them more relevant to the season (instead of the averages for the year).
I've also noticed Google is quicker to update its suggestions on suddenly "competitive" or highly searched phrases.
For example, if a person becomes newsworthy, you could get the top two search phrases from the last, say, 48 hours about that person. I think it used to be based on the averages for a week or two if not longer.
p/g