Forum Moderators: Robert Charlton & goodroi
All the pages except contact, about and main page are supplemental.
e.g.
-what steps you have taken to eliminate duplicate content
-how have you arranged your crawlability
-what is quality of your inbound links
-any deep links
-how long has the site been up
-any changes since it was launched
etc etc
What about the Title / meta description tags? Are they unique too?
Of course.
In discussing the Supplemental Index, I haven't noticed anyone talk about freshness for a while.
Hmmm now I've always disagreed with that. Many of the highest ranking pages in this type of industry (advice/articles) are the oldest and unchanged. What use is a news archive that changes each article every day? ;) I have only found the freshness helps with my commercial websites (i.e. products/services). As many people have pointed out in many threads, Google treats different industries/topics differently ... I would go a step further to say that even the keywords searched for dictate the "type" of search/algorithm Google uses.
internetheaven - a few more details will help your question to be answered:
-what steps you have taken to eliminate duplicate content
Yes, I checked that out, there are a few copies of my articles about the place with scrapers but only a few and none of them have backlinks/PR/etc.
-how have you arranged your crawlability
-what is quality of your inbound links
-any deep links
-how long has the site been up
-any changes since it was launched
The main thing I did (and don't know if this is what helped):
1) Reduced the boilerplate template code to a minimum.
2) Made sure the content was as high as possible in the html.
3) Decreased the common internal nav links - especially on lower-content pages.
4) 'No indexed' a handful of lower-content pages.
5) made sure each title and description were unique
6) got rid of Keywords Meta tag on most pages (I suspect this had no effect).
All of these things were learned here :) thanks guys.
Are the supplementals the latest pages or old pages that have been replaced?
Do you always get supplemental on the results, or does that depend on the search query? [ ie is the site supplemental for all results ]
Has the site tanked at any point which you can relate to an event?
Crawlability - you say you have several revisions of your site. Have you checked the index paths are consistant with e.g. "/"?
How have you organised the navigation? Is it heirarchal and running down through theme levels?
Do you have any "boilerplates"? ... see Adam Lasnik's recent response on duplicate content.
Do you have the content repeated on more than one page of your site?
Can you reveal anything more about the history of major events on your site?
sailorjwd - Were your results affected by these supplementals, and did the changes fix this? Are you able to indicate how the recovery worked it's way through?
[edited by: Whitey at 12:11 am (utc) on Dec. 31, 2006]
Also - for those pages that are Supplemental, how many links are there to those pages? Are they linked to fom 1, 2, 3 pages or more? And what PR shows for the pages that are linking to those Supplemental pages?
It took me a year or more to get rid of the supplementals. I haven't had supplementals for the last 10-12 months.
If a supplemental pops up in the future I would quickly no index it and/or block it with robot.txt and try to fix the problem, and create a new page.
I'm pretty sure my supplemental problem was a combination of a huge template (frontpage) and pages with too little content. followed by some dup descriptions.
My subjective sense is that if we had the ratio of Suppementals to all URLs from last summer, and could compare that to the same ratio in present time, we would see a jump of major significance. Couple that with recent statements from Matt Cutts and Adam Lasnik recently to the effect that Supplementals aren't always "bad" in every case, and I'm beginning to wonder what's up -- what we think we know may no longer be the whole story.
recent statements from Matt Cutts and Adam Lasnik recently to the effect that Supplementals aren't always "bad" in every case,
Isn't that supported by the same page showing in a site: search as supplemental, but in a keyword search as non-supplemental (and rank well), both instances with the same cache date?
At any rate, wouldn't the situation I described make it pretty hard to come up with a count of, or ratio for supplemental pages?
I wonder if a page might be shown as supplemental based on a comaprison to other pages on the same site, but not necessarily when compared to pages on other sites.
About supplemental results, my site went from supplemental results then to main index and back to supplemental results all within 24 hours (I searched site:www.mysite.com).
In addition, the cache was 12/28/06 and now it is back to early November on all my pages. Why?
Oh well, as I said before, the Google dance continues even though the terms have changed. I hope fewer and fewer people will be willing to dance. It is time for another engine to end this ridiculous Google dominating status.
[edited by: JayDev at 4:27 am (utc) on Dec. 31, 2006]
wouldn't the situation I described make it pretty hard to come up with a count of, or ratio for supplemental pages?
You said it -- that's why I wrote that I have a subjective sense. I cannot conceive of a way to nail down such a ratio.
Here's some more detail that prompts me to suggest that something new is afoot with supplementals.
1) I'm looking at a 1400 page client site that had maybe 20 supplemental urls last spring. Now it has 450. Most of these urls 1) would make very weak entry pages into the site, and 2) suffer from some titles and descriptions that are similar to other pages (10 page breakdowns of long articles, e.g.).
One more really strange observation -- for the snippet of these multi-page supplemental articles, Google is using lines 2 and 3 of a 4 line footer. This is for pages that hold 1,000 words and up.
2) On another site that offers many pdf downloads, only those pdf files are supplemental -- and all of the pdfs are supplemental.
3) On a third site, a previous version of the site has been kept online, linked from the new home page as the "archives". Many deep backlinks go to the archive section -- and yet over the past two years, more and more of those urls are now tagged as Supplemental. It's now up to maybe 60%
All of this looks to me like there may be some new criteria for tagging urls as supplemental, but I can't quite get a handle on it yet.
I should add that these newly supplemental pages are getting some long tail search traffic, and that their parent websites are thriving now even more than a year ago. In short, the new supplemental assignments seem not to be hurting at all, and may even be part of bringing in a better quality of Google traffic.
the cache was 12/28/06 and now it is back to early November on all my pages. Why?
1. different data centers?
2. one of those Google mysteries that only we notice because we watch so closely?
Why does one of our sites show supplementals and old cached pages, yet have new caches of the same page:
Site 1 - it crawls around 9,000 pages a day all dupe content fixed, but lot's of sub pages with addditional content, maybe confused with the leading article page. btw - lot's of quality deep links and a PR 5/6
Site 2 is mostly clean with no supplementals and most pages indexed, and no sub pages of the main article, with few inbound links.
Why is Google picking on 1 and not 2. It may strengthen your view on your first example.
My hunch is that it's creating internal duplicate content and confusion with the sub pages - just wanted to run it past you. btw - we're in the middle of an experiment on this network to remove the sub pages from indexing completely.
Marcia
PR dead end pagesWhat is convincing you that this is an issue?
I'm seeing some good SEO's putting outbound links to quality sites and ranking well - any definitive/evidenced connection?
[edited by: Whitey at 7:15 am (utc) on Dec. 31, 2006]
To remain with the 1000 page website as an example: a stray page is one which receives less than one percent of internal links. If you have a website of 1000 pages and less than 10 pages refer to that
page in an internal link, how important could that page be? Drop: I call it a stray. I think this is the minimum new standard.
My sites are undergoing the same experience tedster described in his last post. tedster, by the way, thank you for your participating in this thread.
My sense from observation is that Google is moving to reward websites with wider, more richly linked* footprints as opposed to the narrow footprint site where there is a clear set of home pages on top of 90% secondary support pages of low quality (stray).
In the opening example for this thread, internetheaven described a site top heavy to the home page supported by mostly stray pages. And so Google's latest filtering is following my observations and applying it to internetheaven's example.
For what it is worth, another observation based on IH's example is that 200 pages is the new entry ticket for relevance in Google's world. If a new site does not present 200 pages of quality content and make a wide, not narrow footprint, I don't believe it has a shot of cracking into its market.
*more richly linked per page internally and externally
"For what it is worth, another observation based on IH's example is that 200 pages is the new entry ticket for relevance in Google's world. If a new site does not present 200 pages of quality content and make a wide, not narrow footprint, I don't believe it has a shot of cracking into its market".
If this is the case Google is finished. What most people are searching for is detailed information, something they are not going to find on a "shotgun site". When I search for product or service the last thing I want is to have to search through 200+ pages of dreck to find the nugget I'm looking for. Content is one thing, but throwing up thousands of pages on a topic that could be covered in 50 and giving them preference could only have one reason.....adsense.
Now after reading Texasville's post, I've been trying to figure out this whole base href tag thing! If I'm using all relative links in on my site, is not having the base href tag really going to make a difference? in the SERPs? Thx in advance!
Also another of my sites went completely supplemental (for a long time it wasn't supplementals despite having very similar structure/content to the site that went supplemental).
This points to the "supplementals are increasing" theory.
I just noticed that I accidently copied the 'borders' folder for frontpage onto one of my website folders.
Google indexed the bottom,right,left.htm files.
All are supplemental.
Here are the charactistics of these border files:
1) contain little content
2) are obviously duplicated in many pages
3) have no links pointing to them
4) they have links going out
Hi,
For more than a month now the site: operator returns only supplemental results for my site. the site is pretty big (more than 50K of unique pages), and more than 4 years old.
This is quite bothering since the internal Google's search box
in my site returns search results from the supplemental pages.
the thing is that my ranking was not harmed at all in the regular web search.
did someone had similar a problem? is it the famous "site operator is broken"?
Thanks
[edited by: tedster at 6:06 pm (utc) on Jan. 1, 2007]
I moved your message here because, as I think we can see, your observations are very much on this topic, and they might help us understand better what is going on. You are definitely not alone.
One particular thing you wrote really jumps out at me: "my ranking was not harmed at all in the regular web search."
Does you mean that on regular search, your rankings also do not show the green "Supplemental Results" tag -- but in your Google-powered site search, those same URLs do show as supplemental? If I have that right, do both places have the same cache date?