Forum Moderators: Robert Charlton & goodroi

Message Too Old, No Replies

My urls only show after clicking on "omittted results"

         

dibbern2

4:21 pm on Oct 1, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



On Sept 17 my main site got hit with a big problem in Google serps. I say problem instead of penalty because I can't figure if its a penalty or something else. I've read through the lib here, but haven't found clues to what might have happened to my site.

Here's a snapshot of the problem:
-This is a site that organizes info on 4 topics by state. Think of it as 50 directories with content like *Alabama Rutabaga Farms* where the dir is named Alabama Rutabaga Farms and the title/content on the index page in each dir is Alabama Rutabaga Farms.

-All content is about 4-5 years old, gets updated about once a year, and has normally ranked between 3-12 for the main 3-word term, depending on the competition for the term. Secondary pages in each dir have ranked very well for less competitive terms.

-A search today for *Alabama Rutabaga Farms* would not find my Alabama index page at all, until I select "repeat the search with the omitted results included". After that I'll find my secondary pages scattered through-out the serps, with one or two in 20 position, several more in the 70's, and a bunch somewhere south of 200. The dir index page will still not show up.

-A site: command returns all pages, usually with the index page listed first. My GWT account is showing no major issues, except a big fall-off in visits by the G-bot.

Any ideas as to what direction I might look in to? Is it a penalty? Thanks in advance for the excellent help here.

jdMorgan

5:18 pm on Oct 1, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Are you getting the self-explanatory mesage
In order to show you the most relevant results, we have omitted some entries very similar to the [number] already displayed.
If you like, you can repeat the search with the omitted results included.

or is it something else?

If your site duplicates content available elsewhere, then your pages have probably gone supplemental. This might be because your site has been scraped and duplicated elsewhere, or because it's been made available through a site-proxying "service" -- likely with someone else's ads plastered all over it. Or, if you paid someone else to write the content, they may have copied it from somewhere else (not very likely, from your description of your site as being fairly mature).

You'd do well to carefully check the sites that appear for your search terms, and see if any of them have duplicated your content. Also, try searching for long phrases from your pages, enclose them in quotes, and see what turns up. This helps reveal problems where the scraper sites don't rank very highly, but there are several of them.

Jim

dibbern2

5:29 pm on Oct 1, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



jd,
I get the self-explanatory message you show. My pages have been scraped or made available through site-proxying for years; I don't know how to fight it. BUT, ranking was fine before, even with these problems. Down about 50% now.

If you really think I have gone supplemental, I'll go back to that lib thread. But it seems kind of hard these days to tell what's supplemental.

jdMorgan

5:44 pm on Oct 1, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



We also had some threads here about fighting off site-proxies and scrapers. If you haven't do so, I think you need to address these issues now.

I used "supplemental" in the old meaning; I don't know what it really means any more, either... :)

Jim

g1smd

7:29 pm on Oct 1, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



I'd guess that something has happened (and that "something" may be very difficult to detect) that has caused Google to re-evaluate which site is the "real" one.

However, I have always wondered what might happen when a site that has been copied updates its content, but not enough to escape being a duplicate of the copies, and now the copies are all "older" in Google's eyes. I could see it all going pear-shaped for the site that was copied.

dibbern2

2:53 am on Oct 2, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



g1smd: oh no. I never even guessed of that. Thanks, but that's got me really worried.

dibbern2

3:31 am on Oct 2, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I have been updating these pages, and content is changing by about 10%. Not a lot, but I was hoping this would push G into reranking. BTW, spider activity is up by a large factor.

Robert Charlton

6:03 am on Oct 2, 2008 (gmt 0)

WebmasterWorld Administrator 10+ Year Member Top Contributors Of The Month



However, I have always wondered what might happen when a site that has been copied updates its content, but not enough to escape being a duplicate of the copies, and now the copies are all "older" in Google's eyes.

I've wondered about this myself. It's a worry that kept me for a while from fixing a particular -950 page that was also a page that was scraped to death. Ultimately, we didn't disappear when I changed it, but we also had higher PR than the scrapers.

dibbern2 - Generally, if it is scraping and a page does disappear, Google manages to figure it out in a few days to a week. Google is much better at sorting out dupes, in my experience, than the other engines... but there's a lurking worry that they are not perfect at it.

I recommend you also check your pages with Copyscape, which will detect the kind of scraping that uses bits and chunks of your page. I've noticed some ranking drops in pages that have experienced this kind of scraping, but I haven't established cause and effect. The drops could be completely coincidental.

We also had some threads here about fighting off site-proxies and scrapers.

A basic discussion on proxies is available in our Hot Topics area [webmasterworld.com], which is always pinned to the top of the Google Search forum's index page. Note the Defending Your Rankings section, and this thread...

Proxy Server URLs Can Hijack Your Google Ranking - how to defend?
[webmasterworld.com...]

There's also a discussion on scraped and stolen content in that section.

dibbern2

1:39 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



October 20 update:

My sites have returned, often in the Goog top 5 for 2+3 word searches.

I don't know if this is simply a tweaking of the algo dials, but I doubt it. In all respects my problems seem to have been a penalty.

I undertook a few steps in internal linking that affected each page, and perhaps that was the answer. If anyone is interested, I'd be glad to share what I think was the problem.

tedster

2:29 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



If you can explain the situation without giving specifics that would identify your website, I'm very interested. The "omitted results" phenomenon has almost always been a duplicate (near duplicate) situation in my experience. So if you are seeing something that implicates internal links, I'm all ears.

dibbern2

3:49 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Tedster,
I can't imagine how a duplicate situation would be the problem. My content -before penalty, during penalty, and now- has not changed. Unless, and I think this would be improbable, there was a duplicate "threshold" setting that I crossed over, and now its been rolled back. But that would assume that duplicate content is now okay.

My site deals with health care. Among other content, there are three large directories, each devoted to a specific issue of healthcare. As these grew in content, I thought 'why not link them to each other, since they all appeal to my audience'. To illustrate with some meaningless subject, think of (a) colds (b) flu, (c) sinus problems. Its obvious how I thought my users would find each topic of equal interest.

So during July and August I started adding links to and from each page in each directory to its partners in the other; i.e., cold symptoms <<>>flu symptoms<<>>sinus problem symptoms; and so on, in a round-robin.

On Sept. 17 I went in to penalty territory. My standard measure, a three word search term (example: flu treatment in Arizona) went from about #8 to nowhere, except for searches that contained "repeat the search..." and then, I'd get some obscure page listed in about #300.

My Yahoo rankings actually went up. Long tails at G kept my site alive; I kept about 30% of my old G traffic.

On October 10, I started to remove all those cross links, got about 50% finished. And that takes me today.

I did NOT: modify titles, metas, content (except links) or design factors. Each page update was submitted via GWT.

Don't know if this helps anyone, buit I hope it does. Watch out for cross linking, even though it does make sense to your subjects.

tedster

4:22 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Was the cross-linking done with anchor text?

dibbern2

4:42 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Yes. Text was semi-repetitious , ie, Arizona Flu Treatment, Cold Treatment in Arizona, Sinus Treatment Centers in Arizona. (not the real topics, of course)

There was only one subset of cross links for each locale, but there were several hundred locales.

Also: standard upward links to home for each topic. I did not change these.

dibbern2

4:44 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



added note: I had tried to avoid 100% duplicate word usage in anchor text. I learned sometime ago to not use the blue widgets/red widgets/green widgets scheme.

tedster

5:29 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



One last exploration here - did the pages you were linking to have low PR? ...and did they show up on searches for text in the regular content, or only for words in the Title and Description?

dibbern2

5:48 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Assuming you meant pre-penalty.

Yes, low PR. Usually a 2, sometime a 1, the odd 3.

Ranked well for title/description phrase, and many, many long tails from the content.

BTW, if you can make something out of all this, hope you'll post it.

outland88

6:36 am on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



The key is reverting back to the older copy as quickly as you did Dibbern2. You must get crawled rather quickly. The problem is if you forget the links you developed. What you mention does seem to trip some threshold. I think it has something to do with what tedster is driving at.

I too have wondered many a time about what jdmorgan and g1smd have said.

Social sites can hide copied content in thumbnails and lengthy pages that don't show up in phrase searches. I haven’t seen any scrapers effect rankings but I did file DMCA’s against two social sites and had a site soar in the rankings. You stumble on the copiers just by accident.

tedster

6:32 pm on Oct 21, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



By including lots of similar anchor text on many pages, especially pages that have the kind of partial indexing that we've seen for low PR urls in the supplemental partition, I think you may have blurred the relevance signals of the individual pages too much.

iInventedtheinternet

10:08 pm on Oct 21, 2008 (gmt 0)

10+ Year Member



Tedster. Can you explain this a bit more?

I'm having issues with low ranked pages not getting indexed.

I have lots of search result pages on our general listing site. Broken up by cities and then categories and sub-categories. On those category and sub category search result pages there are 25 results with links to individual listing pages, the anchor text of which is unique and derived from unique page titles. But underneath each of those links of unique anchor text we've also included breadcrumbs for each of the 25 results. So as an example:

TOTALLY UNIQUE PAGE TITLE ANCHOR TEXT LINK
whatever city the posting is from anchor text link >> whatever category the posting is from anchor text link >> whatever sub-category the posting is from anchor text link

TOTALLY UNIQUE PAGE TITLE ANCHOR TEXT LINK
whatever city the posting is from anchor text link >> whatever category the posting is from anchor text link >> whatever sub-category the posting is from anchor text link

TOTALLY UNIQUE PAGE TITLE ANCHOR TEXT LINK
whatever city the posting is from anchor text link >> whatever category the posting is from anchor text link >> whatever sub-category the posting is from anchor text link

This repeats 25 times down the page and the pattern repeats itself around the site. Make sense?

Thoughts?

iInventedtheinternet

10:20 pm on Oct 21, 2008 (gmt 0)

10+ Year Member



OK Let me know your thoughts. In the meantime I've turned those links into text for the time being, I'll report back results.

tedster

1:30 am on Oct 22, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Anchor text is an on-page factor as well as an off-page factor for the urls that the links point to. This is one of the reasons that hover menus and other large menus cause what I've called a "mega menu" problem [webmasterworld.com].

Essentially, that means that many pages begin to show similar relevance signals instead of each page being focused. It's another good reason to limit total links on any page.

iInventedtheinternet

1:40 pm on Oct 22, 2008 (gmt 0)

10+ Year Member



tedster,

So in the above example, which pages might be effected? Just those result pages, the pages that the repeated links point to or maybe even the individual pages with the unique URL's. Maybe all?

dibbern2

4:29 pm on Oct 22, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



iinvent:
From my experience I'd say those links you describe are pure toxic. Just read the thread... the uniqueness of the anchor text will not save your butt. IMHO, you're seeing the start of the problem already with your non-indexing.

iInventedtheinternet

5:34 pm on Oct 22, 2008 (gmt 0)

10+ Year Member



dibbern.. thanks for the response!

OK so if I understand you correctly.. just having them could really screw things up for the whole site.

However I do need to have the result pages with links to each individual post.. otherwise how else can people browse and search. What I've now done, as I mentioned above is to convert from links to just text the links underneath each unique post link.. these:

"whatever city the posting is from anchor text link >> whatever category the posting is from anchor text link >> whatever sub-category the posting is from anchor text link "

In your opinion, would that satisfy you or would you do something else? Feel free to suggest even random unfounded ideas.. the discussion is helpful either way.

dibbern2

1:30 am on Oct 23, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



iInvent: I'm in the middle of bringing my own site back from this penalty, so I haven't much time. My advice is to study the thread, and move forward. Remember, what seems logical to you or what you think your audience would appreciate in linking is not automatically the best scheme when you consider the penalty box possibilities.

iInventedtheinternet

1:16 pm on Oct 23, 2008 (gmt 0)

10+ Year Member



OK no worries dibbern. I'll report back here any results after a week or so. Hopefully I have some good news and we can really expand on this theory.