Welcome to WebmasterWorld Guest from 18.104.22.168
It looks like we are in the same situation.
Our home page which has ranked on the first page, generally 6-9, for the last few years has dropped to #131. Several internal pages that ranked #1 o #2 have dropped as well to #90-#100. Strange thing is that we still have a few pages ranking #1 for their specific keywords.
No link farms or other black hat tactics. The only thing that I can think of is that we are over optimized. Keyword placement in meta tags, h1, and text. While we never intended to keyword stuff, maybe we have tripped some sort of filter?
How do we confirm this or where do we start?
We had a similar situation on Dec. 17. All of our listing just about disappeared or where buried, then within 2 days everything started to come back, and with higher rankings. We didn't make any changes to the site.
This most recent incident occurred on 2/3, and we haven't seen any progress with our pages returning to normal in the rankings.
I know many say they are not over-optimized for those keywords, but that is according to past guidelines. Perhaps G has lowered acceptable kw density for competitive phrases, in attempt to combat spammers. Are you saying those keywords don't show up at all on the affected pages? How much?
Look at similar sites (similar backlink, pr, etc) that now rank high in serps for those keywords - how much do the keywords appear on those pages?
into google my site shows up at # 43 in the listings.
If I type
It shows a mathematical site first and then one page of my site second. If I do a
It shows all of my pages..
Also, a keyword that I used to rank # 1 for I now rank # 40. I am not ranking for any of my keywords anymore. Don't know what this falls under but I am totally baffled.
[edited by: tedster at 9:18 pm (utc) on Feb. 7, 2007]
[edit reason] use example.net [/edit]
My site came back this morning, after suffering the symptoms mentioned here for 4 days. I suspected over-optimizing or keyword stuffing, so I reduced my keyword density. Googlebot came back to visit after a 2 day hiatus. Today, the main page caches is updated and I'm back in the SERPS just like I was before. Did the changes I made fix it? who knows.
Had that experience too. Changed the format of my homepage 2 months ago, which affected the KW density. Site fell from page 1 to page 2. Immediately reverted back to old format and went back to page 1!
Looking at the serps for my niche though, I am convinced (in my own mind anyway) that Goog periodically "abducts" the #1 site for a couple of days, send them off into the lab for poking and prodding (filters), then if they pass the tests, they come back at their regular place.
Then again... what do I know... sigh!
What about keyword density in the title, hrefs, and H1 tags?
Title - The the titles of these pages are all the same:
"Our Site Name: Red Widget"
hrefs - The hrefs that link to these pages are:
There are no other internal hrefs for these pages.
H1's - We have no H1 tags on the entire site. Nor do we bold or italicize the keywords, although at the top the keyword "Red Widget" is in a larger font than on the rest of the page, but this is purely aesthetic and has nothing to do with SEO.
We now rank in the hundreds for "Red Widget" but if you use a 3rd keyword and type "Red Widget Accessories" or "Large Red Widget" (for example) we're right back at #1.
My site has a home brew CMS which I have been optimizing like crazy for the past few months... perhaps I tripped a filter or something, Google may have seen too many changes in title and description tags, content etc. over a short period of time. I may have over optimized. I decided during the down time that I might as well take the plunge and convert all my URL's to friendly static style as advised by others here at Webmaster World... so, going from index.asp?w=30 to /widget30.htm with 301 redirects may put me back into the darkness again for a while, but hopefully it will pay off before too long with better rankings and stability. Has anybody else ever done the change from dynamic to static URL's? How long did it take to recover in G's index with same or better traffic?
It is almost as if someone at Google hand picked the exact high-traffic key phrases and put them into penalty or something.
Why "hand picked"? Couldn't they be looking at searches for high traffic phrases and filtering/penalizing pages that are too highly optimized for those phrases?
I didn't really mean hand picked literally, but just very well targeted and focused on specifically my target phrases. Very "AI" ish!
But really, I'm just suggesting perhaps they aren't using this "filter" for less competitive phrases, so you still rank for those...
Has anybody else ever done the change from dynamic to static URL's? How long did it take to recover in G's index with same or better traffic?
yes one of my client who was ranking #1 and #2 position for last couple of years has done some major changes in site layout and urls since last dec 2006 and it seems like goole has hurt him badly what we can do is wait and watch if we can come back or we have to rollback or we have to find what is gone wrong. As per me everything was very much as per google guidelines and no spamy technique or no deceiving at all. Still my client is hurt he is no where in SERPs for any keyword.
Please if anyone can know if it is just to my client for his global changes or it is google who is doing some algo changes
Google is changing almost daily now, vikram, so it's hard to answer you more exactly. I would not try to back out, however, I'd just make the new way work - I know it can.
Netmeg - I'm still thinking the site command is not working right, or else they changed it. Most of my sites are display as filtered content right now (mine and my client's) and not as supplemental, and many of them are as high or higher than they've ever been in the SERPs.
My hunch is that it's intended and sites are at risk with the "duplicate content" dials being tweaked . I think it's no accident - be wary and prepare [ IMO ] .
g1smd has made some early comments: [webmasterworld.com...]
I think some special focus seperated out from the above on the specifics of what the site:tool might be trying to tell us is urgently required for some sites to avoid the fall, even if you previously believed that you had cracked the duplicate content filter.
At it's lowest it went down to 1-3 of 21. As of a few hours ago it's 1-10 of 960. The first three results are the ones that were showing before and are 3 major pages on this site. All of the results after that are supplemental however doing a regular search brings up many pages that are not supplemental. It appears at the moment that the site operator is only returning my 3 main pages plus supplementals. Pages that are not supplemental are not listed, at least as far as I have looked.