Forum Moderators: Robert Charlton & goodroi
My theory is that since we use source markers on many pages as an easy way to internally recognize the traffic to a specific page, whether the traffic came from email, specific advertisment placement throughout the site, etc... But i feel each of these src markers is being treated as seperate and unique pages - is that true? If so, how do we fix this issue?
Thanks in advance for anyone's input on this...
Vic
As a rule of thumb, most click-trail tracking is best done with cookies. You only want Google to see one url for one bit of content, and not alternate urls that get the same thing. You can use robots.txt, robots meta tags and so on to limit what you want Gbot to see.
ie...
www.domain.com/page1.html?src=blah
and have the robots.txt tell the spiders that the page name really should be...
www.domain.com/page1.html
is that possible with the robots or any other method? i dont want them to exclude all of these pages because they are heavily linked and some have a PR of 6 while others have 4s. My main domain has a PR of 8 and I want to be sure to take full advantage of this.
Thanks for your help...
Vic
$userAgent = getenv("HTTP_USER_AGENT");
$queryString = getenv("QUERY_STRING");
$domain = getenv("HTTP_HOST");
if (stristr($userAgent, "Googlebot")) {
$isSpider = TRUE;
}
if ($isSpider AND !empty($queryString)) {
header("HTTP/1.1 301");
header("Location: [$domain$PHP_SELF");...]
exit;
}
This PHP code will redirect Googlebot to the clean URL w/o ?src=anything, while all other user agents are tracked.
"Don't use "&id=" as a parameter in your URLs, as we don't include these pages in our index."[google.com...]
You'd just want to be sure that Google crawls some version of those urls.