Forum Moderators: Robert Charlton & goodroi
I thought I’d start a list...
Domain
- Age of Domain
- History of domain
- KWs in domain name
- Sub domain or root domain?
- TLD of Domain
- IP address of domain
- Location of IP address / Server
Architecture
- HTML structure
- Use of Headers tags
- URL path
- Use of external CSS / JS files
Content
- Keyword density of page
- Keyword in Title Tag
- Keyword in Meta Description (Not Meta Keywords)
- Keyword in KW in header tags (H1, H2 etc)
- Keyword in body text
- Freshness of Content
Per Inbound Link
- Quality of website linking in
- Quality of web page linking in
- Age of website
- Age of web page
- Relevancy of page’s content
- Location of link (Footer, Navigation, Body text)
- Anchor text if link
- Title attribute of link
- Alt tag of images linking
- Country specific TLD domain
- Authority TLD (.edu, .gov)
- Location of server
- Authority Link (CNN, BBC, etc)
Cluster of Links
- Uniqueness of Class C address.
Internal Cross Linking
- No of internal links to page
- Location of link on page
- Anchor text of FIRST text link (Bruce Clay’s point at PubCon)
Penalties
- Over Optimisation
- Purchasing Links
- Selling Links
- Comment Spamming
- Cloaking
- Hidden Text
- Duplicate Content
- Keyword stuffing
- Manual penalties
- Sandbox effect (Probably the same as age of domain)
Miscellaneous
- JavaScript Links
- No Follow Links
Pending
- Performance / Load of a website
- Speed of JS
Misconceptions
- XML Sitemap (Aids the crawler but doesn’t help rankings)
- PageRank (General Indicator of page’s performance)
I think some of those may be the final effect of specific factors, rather than being factors on their own. Here are some ideas I've had:
1. History of past penalties for this domain
2. History of past penalties for this owner
3. Semantic information (phrase-based indexing and co-occurring phrase indicators)
4. Taxonomy flag for general category (transactional, informational, navigational)
5. Taxonomy flag for freshness
6. Taxonomy flag for market niche
Non-Link word count:
How many words on a page are not links? More words that are not links is a general indication of more "real" content on a page.
I believe these are a factor, or at least looking at these two together is a good proxy of things that are indicators, in terms of "is there content on this page?" evaluation.
A page that's all links is bad.
Incorrect, I have many trade widget pages like this and they all rank #1...caveat - for MY widget industry:-)
@ peterdaly - I know where you're coming from however specialised widget sites can do extremely well with these.
Currently I don't see evidence of keyword in H tags as a ranking factor. Those days are behind us
Probably some of the stuff left over from the old days is still in usage:
- Word emphasis (bold, italics, underline)
- positioning of most relevant words on page
Since H-tags often are bold, italic or underlined as well as prominently placed I would be willing to say that they have some relevance - not as H-tags, but because of style and position. Probably in the 180-200 position on our list of top factors.
...but I do agree with tedster, the magic is gone.
I am not sure if the following one applies though.
Per Inbound Link
- Authority TLD (.edu, .gov)
From Google groups Oct 2008 [groups.google.com]
TylerDee, TX: Are .gov and .edu back links still considered more "link juice" than the common back link?Matt Cutts: This is a common misconception--you don't get any PageRank boost from having an .edu link or .gov link automatically.
If you get an .edu link and no one is linking to that .edu page, you're not going to get any PageRank at all because that .edu page doesn't have any PageRank.JohnMu: We generally treat all links the same - be it from .gov or .edu or .info sites.
.gov and .edu sites generally have been around awhile and over time gained trust, so I guess that can make them seem more authoritative.. But on the flip side, authority sites tend to get targeted by spammers, .edu sites especially get hammered all the time...
Try this search for spammer-targeted keywords:
site:.edu/ keyword
What affect that has on trust is anyone's guess.
IMHO
Cheers
James
[edited by: Robert_Charlton at 6:21 pm (utc) on Nov. 24, 2009]
[edit reason] removed specific search [/edit]
Matt Cutts: This is a common misconception--you don't get any PageRank boost from having an .edu link or .gov link automatically.
If you get an .edu link and no one is linking to that .edu page, you're not going to get any PageRank at all because that .edu page doesn't have any PageRank.
Matt said this in 2008 and he is talking about pagerank - a factor I tend to pay less and less attention to.
In my experience all .gov/.edu sites were not created equal, but I am a believer in the trust/authority factor.
Matt said this in 2008 and he is talking about pagerank - a factor I tend to pay less and less attention to.
Excuse my ignorance, but isn't this discussion about PageRank and the factors that determine it? and I don't mean "toolbar pagerank" (a factor I pay zero attention to), but the algorithm Google use to rank a page in the SERP's.
I am a believer in the trust/authority factor.
I also believe in the trust/authority factor. I just don't believe TLD's have anything to do with it.
Cheers
James
Lots of good points I'd like to comment on.
@Hissingsid
who says speed = "- Performance / Load of a website "
@peterdaly
Link density is something I haven't considered. I'm going to read up on this.
@tedster / gn_wendy
I think you’re correct about the H tags. My 2 cents is that these tags have devalued in the past 12 months but I believe they carry more weight than standard text.
There has been very few additions on the link building factors.
Thanks
Donal
...PageRank and the factors that determine it...
Point taken.
I also believe in the trust/authority factor. I just don't believe TLD's have anything to do with it.
IMO .edu/.gov site's have an easier time of getting to a higher level of trust/authority than other TLD websites. If this has to do with the type of website or the TLD is a bit beside the point ... if the shoe fits.
This was brought up by several speakers at PubCon and Matt Cutts confirmed Google might go ahead with this in 2010.
I've listened and he mentioned speed and fast but not "Performance / Load of a website". The point I'm making is we don't know what Google means by speed yet. Or rather we don't know the most economic ways to make Google think our site is fast yet.
Cheers
Sid
P.S. By visitor behavior I DO NOT specifically mean 'Bounce rate". Instead, I'm talking about behaviors like bookmarking and return visits.
In the "Penalties for the domain" list, I'd add CSS processing (excuse my english) for "display:none" CSS and maybe others that make the text not show up to the user.
P.S. By visitor behavior I DO NOT specifically mean 'Bounce rate". Instead, I'm talking about behaviors like bookmarking and return visits.
I'd guess that (along with "freshness" and and number of other factors), visitor behavior would be judged in context--i.e., in relation to similar pages or sites.
Take forums:
- A good pregnancy forum or rugby fans' forum should have a lot of repeat visits, because visitors are interested in the topic and can be expected to come back.
- A good Windows support forum should have fewer repeat visits, because once the visitor's problem is solved ("Help! I can't find the other computer on my home network!"), he shouldn't need to come back until the next time he has a problem.
In the "Penalties for the domain" list, I'd add CSS processing (excuse my english) for "display:none" CSS and maybe others that make the text not show up to the user.
How would you suggest this applies to image swaps, especially when the image and the text seen by Googlebot are identical and how do they tell the difference or know when to apply a penalty? Or, are you suggesting 'display:none;' is an automatic penalty?
It's definitely not (in my experience) a site wide penalty to use an image swap on a page, which requires 'display:none;'. (Actually, in my experience, it's not even a single page penalty to use an image swap...)
[edited by: TheMadScientist at 3:45 pm (utc) on Nov. 24, 2009]
Not sure if Google is capable of doing this kind of CSS processing (as deep and as complex) but pretty sure some CSS elements/properties are getting checked automatically, either the css code is embedded on the code or included or on an external file.
...or there is a padding or z-index element which makes this hidden to the user.
Personally, I can see this one, but not the first half of the statement...
I'm saying that if there is a little paragraph or block with links and the display is set to none...
I actually know of well ranking sites where that's absolutely not the case.
* I should note, WRT the preceding, they do make use of AJAX though, so it's not 'invisible' to user without viewing the source code, but the visitor does have to click on a link to get it to display...