You betchya Tallguy!
Report them all you want. From the experience of others on here, if GAd Does decide to kick them off, they'll just be back next week under a new domain.
Adsense is just ONE of the money makers. Many of them are affiliate or other PPC publisers too. You could spend all day identifying them and tracking down their affiliates, etc.
japanese - if a site has a 302 link to yours can't you just deny access to that ip? The link will show as blocked so they would remove it from their page and googlebot can't follow it either.
I have 2 of these type links in my allinurl: and all MY links show as "removed some links which are less relevant" my whole site is less relevant to my url than these 2 damn links?
These arn't bad guys either they are free links on paid inclusion directories (from when I was in the sandbox).
There is no 'remove url' but there is a 'report bad links' so I just blocked them and reported them.
Also I have a looong cgi type one there too - no title no description just this link from about. They display one of my pages in a frame. I was thinking to just put a framebuster on that page.
Hijackers get the traffic from search engines, so we should keep reporting them to major search engines.....to discourage these hijackers.
I assume we all could together and fill up the spam reports :
|The act will be against your principles and you will not be able to sleep knowing you pointed a gun at someone innocent. |
well it just seems we'll have to choose our target carefully. There are plenty of large corporations that maybe innocent in this 302 hijacking issue, but are far from innocent on the grand scale of things.
A little bit of hacktivism to kill 2 birds with one stone perhaps?
"if a site has a 302 link to yours can't you just deny access to that ip?"
O.K. but you will need to deny Google's spiders for this to work. When the googlebot visited the redirecting site they recorded the link to your site as a temporary repository for content belonging to the redirecting site's URL. Later, another spider follows that link TO your site. To prevent the redirecting site from getting credit for your page... you would need to ban googlebot from access... then the redirecting site will get zero. But then again so will you. It's a no win for you. I will repeat this again though it seems to do no good... there is no real difference in representing a third party websites copyrighted work as belonging to a URL you control from actually cutting and reposting it.
On a side note these folks DO sometimes use link checkers to see that their redirect targets are still live. I am of the belief that denying this activity does help as it causes the redirect site owner to believe the link to be bad and they will change it to point at somebody elses site. For this reason I ban the i.p. address blocks of discount web hosts and bargain basement colocation facilities as I come across them. Home cable modem and DSL users that display this type activity get trapped one at a time.
Why is everyone attacking Alexa? Ask Jeeves does it too. About.com does it too. Even Yahoo Directory uses the exact same 302 method to track clicks.
Do a search for site:rds.yahoo.com and you'll see thousands of inurl:somedomain.com. Perform a manual get of the links and you'll find that even Yahoo is using 302 redirects to send visitors to the destination page. Hijacked by Yahoo? Thousands (millions?) of other websites use outbound click-tracking, the majority use some form of redirect/refresh.
The Internet is not the problem.
Google is not the Internet.
All this energy could better be directed at the folks actually responsible for the SERPs. (hint, starts with 'G')
I don't care obout the millions of sites tracking clicks but when I do allinurl on my own site and #1 (even above me) is some fly by night hosting directory listing then that concerns me a little.
I get crawled by google constantly, I would rather see that one listing dissappear since it seems to be stealing my fire somehow.
also I just want to block that one link from that one directory - if that means I have to block a range of IP's belonging to that company then so be it - they don't send me any clicks anyhow.
Perhaps you are correct. If enough people here wish I may lay down a blueprint of how it can be acheived.
We could pump out CGI controlled 302 redirects to say an entire network of a targetted site.
Provide the paths for googlebot to go to the site via a 302 directive. Including paths via search results in Alexa, Yahoo and other search engines and directories who have loopholes in the production of their results by them displaying the address of the result. A website full of links leading to the 302 in all possible ways.
Monitor the results and compare to before we did it to see if an effect had been acheived.
We must first check its inurl: make sure it is clean, then we pump it untill it goes down in rankings.
Big possibility.... Let me know if enough volunteers.
It may also be possible that google is aware of this thread. I deliberately pointed out how easy this trick can be done. They now know that if enough people see this thread then it may play havoc on their system and other websites.
Their engineers would also know I left out a few key factors that causes googlebot to create a duplicate page. It involves their freshbot and deepcrawl which I mentioned in this thread. I am near 100% sure of the conditions that are needed. I lack the skill of a professional blackhat who does it day in day out. You simply keep pumping away at the target relentlessly providing the bots with the linking paths to the redirects. RELENTLESSLY until the site goes down or duplicate pages start to appear.
Don't be surprised if the 302 problem goes away in the near future.
But before we do this to a website, we must all agree to warn the targetted site. Hey, it is not illegal so they should not mind and welcome it. We are only testing all sorts of things like how well our 302 method works and monitoring the number of robots pass through our 302 on the way to their site. I don't think any site will mind it.
Please contact a high ranking site as soon as possible, let them know what we are going to do so that we get on with it as soon as possible. Try calling them by telephone to get their approval to speed things up.
|also I just want to block that one link from that one directory - if that means I have to block a range of IP's belonging to that company then so be it - they don't send me any clicks anyhow. |
you don't seem to be understanding this, You cannot block an "link".
Google goes to the hijacker's site, sees the url that is nothing more than a 302 redirect and sticks it in its database.
later googlebot (from google's IPs) comes and spiders that URL and is redirected to your content. Gbot never realizes that the url is not really where the content is.
please tell me what IP you plan to block? the page that is linking to you? it will do no good. the only Ips that ever access your site are the IPs of people/bots that are viewing your site.
japanese, I've organized that mailing list, we could co-ordinate this through that so that we can use specifics.
I'm one volunteer and i'm an experienced PHP developer who is ready to do some serious coding if need be.
Lets do this, enough talking.
|Google is not the Internet. |
Unfortunately, for about 70% of the people who use the internet, G is.
"use the <base href=""> meta tag on all your pages"
Claus, if you use only absolute referencing in your pages and no relative, should you still install this?
also japanese: how would we manage to get these links on a domain that is already listed in google someone would have to dontate a perfectly good domain name.
now that i think of it i have a few we could use, they'd need some incoming links to get the bots really moving but they'd work.
and who are major google investors with websites we could target?
A VOLUNTEER PLEASE
Somebody please contact a high ranking site that gets about 10,000 unique hits per day and inform them that a group of webmasters are planning to pass a vast number of CGI, GO-PHP META REFRESH and other variant scripts that cause 302 redirects relentlessly to their network of web pages 6 levels deep. It is not against the law and that they have no right to refuse us.
We want to implement our testing within the next few days and that our ultimate goal is to expedite the creation of as many dulicate pages of their website as possible in googles results. We will monitor results and inform them of our findings as we go along. Our main target page will be their index pages within subdomains and folders. But we will not exclude any internal pages because we want to see how many of those we can also cause duplicates of in google's databases.
I will be very interested to see if the website agrees. They already are being 302'd by Yahoo, Alexa etc etc, so it should make no difference to them.
Why do I think they will reply with legal threats! We will see.
We must have the backing of Claus and Brett. If they endorse it. Lets do it.
See if they agree. Hey, there is nothing wrong in causing 302 redirets, right?
what about a mod rewrite? if the referrer is from that site then redirect it.
what about a mod rewrite? if the referrer is from that site then redirect it.
I don't think googlebot passes a referrer entry.
I will lay down the blueprint.
Clause will endorse it with a few modifications.
""""BANG"""" WE DO IT.
I can see it now, just like a hail of bullets leaving a trail of speeding neon lights in the night sky. Thousands of them, maybe millions. Depends on how big the target is.
Googlebot has no referrer. Stargeek made a brilliant suggestion that, should googlebot carry a referrer, then it would be easy to block it.
The idea can possibly work, if googlebot is forced to notify all 302 referrers, then we know when the danger is coming from.
Referrer only when the bot is on its way to you via a 302.
Simple and effective.
I just read where a big dot com got banned by google for doing 302 redirects. Interesting.
You can find the article by searching for "Google denies 302 redirection problem".
[edited by: Emmett at 5:29 am (utc) on Mar. 11, 2005]
|Stargeek made a brilliant suggestion that, should googlebot carry a referrer, then it would be easy to block it. |
they'll never do this tho, it opens the door to too many reverse engineering possibilities for thier algo.
wow yeah thats a big site i'm not sure if this is allowed but here's an important part of the article:
[quote]A Google representative, speaking unofficially, suggested that only one "concrete" example of a domain being hijacked in this manner had been proven to Google engineers. However, he urged the wider webmaster community to post any examples of "domain jacking" that they find to webmaster [at] google.com, with the keyword "canonicalpage" in the e-mail title, so that Google engineers can properly investigate whether there Google feels there is a real charge to answer.[\quote]
the article also mentions a big name "SEO expert" who agrees that this is an issue, perhaps we could get his approval on a campaign.
why not a campaign of e-mails to webmstr @ google with cannonical in the subject line?
I'm not talking about blocking goglebot referrals I'm talking about screening referrals from known problem links and blocking them.
|I'm not talking about blocking goglebot referrals I'm talking about screening referrals from known problem links and blocking them. |
sorry if i was snappy in that post this is causing me headaches (google, not you).
the only way to know where a hit came from is to check the http variable "referrer" normal web browsers "carry referrers" meaning they provide a value that is the last page they were on.
if a normal web browser goes clicks a link from example.com to your site and your site checked the referrer string it would see example.com.
problem with this is that google does not follow links in this way and therefore there is no referrer value to check.
googlebot's traffic from the offending site to your site is the traffic you'd want to block, but it is impossible to tell if google is coming from a redirection or just indexing your site.
Ha what they did is a 302 redirect on themselves.
Mod rewrite domain.com $www.domain.com
as a 302 type redirect and the google duplicate content filter banned them.
Better go check your htaccess files guys this MUST be a 301 or you will get booted out of the SERP's.
I also found googleguys post in another forum that the article refers to.
|(i.e. "site A appears to have duplicate pages from, or is doing a 301/302/whatever to site B, and Google is wrongly picking site A as canonical", with actual values for A and B), I'd be happy to hear them. Drop an email to webmaster [at] google.com with the keyword "canonicalpage" (all as one word) and I'll ask someone to collect the feedback and pass it on to an engineer. |
anyone who has been hijacked start sending then e-mails of known examples.
i'm sure they know of this issue and don't care or can't fix it right now.
all the emails to them in the world won't make it worth it to them to fix it, every time a webmaster looses free traffic google gets an adwords customer.
we have to take it a step further, google won't fix this on thier own.
|I want to know has anyone else experienced a similar situation as mine. All my sites are hosted with GoDaddy by the way. |
Yep. Hosted at GD as well.
I bet that was the reason behind googles rollback.
Big dot com engineer decides to modify htaccess so that non-www redirects to with-www but does not specify 301, default is 302.
Google spiders big dot com twice, once with www and once without so there are 2 copies of every page in googles index.
Googles duplicate content filter bans big dot com.
Big dot com freaks, fires engineer and freaks on google.
Google does rollback on index to restore big dot coms SERP's immediately.
How about bringing down the IRS web site.;)