Forum Moderators: Robert Charlton & goodroi
also, if your not registered, sign up for google webmaster tools...
see if that page is listed as a "sitelink" and you can block it from there...
dont know if that will work, but its worth a look.
1. More PR for the url - an external backlink or two works wonders, or improving the internal link structure to circulate more PR to the page.
2. Unique Title element, relevant to the page.
3. Unique meta description, relevant to the content of the page and concise (no more than 160 total characters.)
If there are any canoncial URL issues [webmasterworld.com] with the domain, get them patched up as much as you can, too. That will also help make the most of your PR.
I assume there are various undisclosed factors, and I also assume disclosed factors change over time, but two other factors that have been mentioned by Google are 1) URL complexity and 2) content staleness.
The Supplemental Index, in its original incarnation at any rate, was not fully searchable. That is, all the content was not analyzed and tagged for relevance, but key factors such as Title and Description always were. So if those two were not unique, the page looked more like something Google wouldn't include in the main search results.
Most common problems to go SPI:
- duplicate pages
- identical title tags and identical description meta tags
- not enought indexible content
- too many parameters in the URL
- the page is not linked to well enough to warrant regular crawling or the page has no incoming links but remains indexed because it used to have links
Unique meta descriptions and title elements have definitely been an issue in borderline or "edge" cases - low PR specifically. I have restored scores of urls to "findable" status by putting attention on just these two factors.
Tedster, you are also changing on-page content, which we do know influences a page's "findable" status.
It's also extremely difficult to isolate factors on an online SEO "experiment" because there are often unanticipated factors at play. For example, tests run to prove that Google dropped the second link when there are two links pointing to the same URL failed because the testers didn't anticipate factors like identical anchor text.
A site's PageRank distribution also shifts like waves over time - it isn't a constant even if you don't lay one finger on a site and even if backlinks to a site remains the same.
all the content was not analyzed and tagged for relevance, but key factors such as Title and Description always were.
Right, but duplicate text isn't an issue as Googlers have often reiterated.
Supplemental index is a way to prioritize crawl order: to pick up fresh content on high authority sites quickly while shelving stale content on untrusted sites for later crawling.
In that process, content freshness, url complexity and PageRank are three factors we know with absolute certaintly which influence a URL's supplemental status.
[edited by: Halfdeck at 4:17 pm (utc) on Dec. 30, 2008]