Welcome to WebmasterWorld Guest from 188.8.131.52
Forum Moderators: open
There is a site, which i can't give any details about, which has 200 backlinks showing, including two PR6 and mostly PR4's. It has no PR7 or higher links.
But it has a PR7.
The only possible explanation besides a glitch (unlikely) is that this site has about 30 (could be many more) inbound PR6, 1500 inbound PR5 (could be twice that) and several thousand inbound PR4 links that come in via cgi redirect, ad servers or other referal method.
Taking a step back, this site *should* be PR7 with all these links going to it.
Google has found a way to give this site the PR it deserves, but it doesn't show us the backlink reasons "why". I believe this is an enormous bit of information with lots of repercussions for sites that offer a lot of cgi links to other sites.
Discussing the PR directories pass on which use jump urls the conesensus seemed to be, yes they do, if their own PR is high enough.
I can confirm to that from a large directory, whose cgi links sometimes even showed up as backlinks in Google.
I'm saying the backlinks shown via the toolbar may only be a very small fraction of the backlinks being actually counted by Google.
I'm saying looking at the backlinks for some particular types of sites is completely useless in analyzing that site.
I'm saying that if you are a site that offers cgi advertising type links to some sort of business entity that you should include in what you charge them some calculation for PR that you are giving them.
And finally, I'm saying that all the concerns people have about getting links to them via cgi "masked" redirects should not concern people at all, because even if you don't see the links listed as backlinks they are still being counted by Google.
joined:Oct 19, 2002
The only *possible* way this could happen, I don't know if it would show up as a backlink or not, is if the redirect URL included a parameter such as "?redirect=www.domain.com".
will all be picked up by google as links, even if your cgi-bin is blocked by a robots.txt
Robots.txt is only advisory. Its still open for any spider to spider it, and you know what some people are like! Are you sure that this was just the case with a rogue spider from an email crawler or the like?
All the pages are greybar and are not optimised and his static homepage is an unoptimised PR3.
I have a client who has no backlinks on his site, yet another site links to him via an clickeable image. I know it cannot be the image that google does not support since I have other images that show up in backlinks on other sites. Both sites have been indexed by google, and both pages have been live for a long time. The incoming link is on the homepage of the other site, and the site does not have many links. The site which hosts the incoming link is only a PR1, so maybe google doe snot track links from PR1 or lower sites? The only thing I can think of is that the target=blank (opening a new window) is not followed by google.
Anyone else have any experience with this. Maybe this should be a new thread.
That makes some sense and is what I was trying to clarify with my earlier post. Thinking about it now, I remember that Yahoo links out in exactly this way, and we know that Google can spider Yahoo with no problems.
I'm not saying googlebot actually downloads/follows the cgi links, just that
googlebot makes an assumption that it is a link to a certain page
based on the fact that the script parameter is an url
I've found that the only way to hide a link for googlebot for 100%
is to use a code in the link (a database key or smth similar), so
that your links look like
(together with blocking your cgi-bin, or at least your link script
in your robots.txt)
I've read other threads here saying you can also do this with
googlebot can follow these links too, so better safe then sorry