Welcome to WebmasterWorld Guest from 22.214.171.124
Forum Moderators: open
Most designers will use relative links, I assume becuase it's easier for them to manage, but will this put off the spiders, particularly Googlebot?
There is a good point to the suggestion Google may blindly follow the link as if it were an external link. If Google liked following external links more than internal, you may have a point. (i doubt there would be an advantage to this though)
The point I wanted to make on the subject is... I think this would slow down your site, as the users PC would likely do a DNS check on every click (to resolve the domain name to the IP). IMO.
imho, http-clients do always calculate an absolute uri from relative links before requesting the linked resource.
i.e. they're able to generate a complete http-request from the relative link, including the protocol, host etc.
if there's a relative link, the client checks for a <base href="..."> element.
if there is no such element, the document's uri is assumed to be the base href.
i don't think googlebot would care about a link being absolute or relative.
In my opinion: while linking into your same domain, use relaitve. keep absolute links only to out-of-domain links. My page is in a subdomain, given as a free service by an important web portal, and my page contains somthings like this:
<a href="/">Visit our hosters</a>
Even this link goes out of my site, its destiny is in the same domain, so i use relative linking.
So do the best for your users, and if you are worried by SEO, think this: If the objective of Google is to rank the page by their quality and relevance, then they will care everything they find is good for the user.
We use such a naming convention to identify the top level, and second level category to generate the left navigation. So this file would be part of
products & services/trading widgets
I just wanted to make sure we're not penalized for using periods and such long filenames.
Does anyone know how Google handles these types of urls
That is not quite correct!
If you go down a directory level (away from your root) the toolbar add-on for Internet Explorer reports a lower PR, so far so good. But this is not the real PR as recorded by Google, it is only an estimate based on your homepage's PR.
So, you don't have to worry if not all pages are at your root. Or the other way around, it does not help at all if you use your naming convention, because for the calculation of the true PR only the link structure is important and not the directory level.
Hope that helps!
NOTE: Google has going about updating it's pr information on a continuous basis yet and shows pr based on the last major update.
I don't like ../../../overthere/page.html links as I get confused where they are pointing. If moving things around (the page that contains the link, that is) then it is easy to forget to edit a link and so it will no longer resolve.
I like /some/folder/deep/down/page.html as I can see where it is pointing and quickly spot a mistake. If the pointing page moves folder the link does not need to be edited.
One point about links that do not contain the domain. You risk having the wrong domain returned for a result. I had a site on a .net free host. It has many backlinks. One person linked to the .com alias and as it happens, the site is listed in Google as a .com URL. It has taken months for Google to see that they are the same site, and list the same backlinks whether you specify .net or .com versions.