Forum Moderators: Robert Charlton & goodroi
This morning while looking to see if there were any extra proxy highjack pages in Google’s index for my site I got a very pleasant shock.
Zero
Nadda
Ziltch
I’ve done nothing to ban and I could not start lookups on my server so its not anything I’ve done.
Has anyone else noticed that the proxy page totals have disappeared could this be down to the authority patent that is being discussed on the October serp’s changes thread?
[webmasterworld.com...]
Anyone else seeing this?
Vimes.
You can also try inurl:example.com where example.com is your domain. Many, but not all, proxy servers will include your domain in the url somewhere.
Still - I'd love to hear other reports. Has anyone seen their previously hijacked positions come back to them?
losing the the 'supplemental' status on these landing pages is a just what i needed, i read both of the patents last night and from what i'm surmising the authoritative document is addressing these proxy issues.
Anyone have any other ideas?
Vimes.
This could be related to what I saw on our sites.
Just that the pages previously hijacked have been removed from the Google index and can't be found at all.
At least for now.
The pages that recently disapeared on our sites were 2-3 y/o with significant links internally and from other quality sites.
They have been hijacked back 10 months ago and mostly came back little by little - although the GWT never showed them as part of the site anymore, they were cached and crawled.
Now our pages are gone again, and the proxy pages are also gone.
Looks like a no brainer for Google, remove both the proxy and hijacked pages just in case.
I still see some proxies with copies of other sites around. So not sure if this is what is really going on.
I'd love to hear from others and see if noticed the same thing or if what we, again, suffer from is totally unrelated.
Having a batch of them disappear is nothing unusual because many of these proxy sites are fly-by-night and are gone before you know it. This is because the value of the typical CGI proxy is only good until the schools and offices have it blocked, then they need a new domain name.
I see the following indexed for one of my sites in the 2 and 3 position for a special page I feed just proxy sites so they don't hijack real content:
www.proxy1.org/nph-page.pl/000000A/http/www.mydomain.com/
www.proxy2.com/asd/nph-proxy.pl/000000A/http/www.mydomain.com
...and a few more proxies further down the page, so yes they're still out there and Google is still indexing them.
[edited by: incrediBILL at 7:03 pm (utc) on Oct. 11, 2007]
Can we go one step further on the discussion? Even though Google is still indexing proxy server pages, is there something new in the algo that keeps them from ranking so easily - at least for common searches? I think the recent relative silence on this previously noisy issue is interesting.
Shame I can't cite an example here.
As I mentioned on Dan's post, one of the main problems now is that the CGI proxy hackers can remove (and are removing) the meta robots noindex tag.
I've been doing some research and I think I came up with a good solution. Here is the basic idea:
1) Use a global shared blacklist of CGI proxies being used to hijack
2) Use detection code that checks every visitor against that BL and blocks CGI hijackers
3) Set up traps that catch the CGI proxies and update the blacklist
The good news is that the project honeypot http:BL already has everything we need. We only need to work with them to create honeypot (traps) designed to catch CGI proxy hijackers.
One simple way to create such trap is to have the honeypot page generate some encoded text with the IP (and other identification information) of the attacker. The page must be blocked by robots.tx and include the relevant meta robots tag noindex, nofollow. If there is a page that comes up for the encoded text, then it has to be a CGI proxy.
[edited by: tedster at 3:01 pm (utc) on Oct. 12, 2007]
I don't see any significant changes on the larger scale at all. The good people at Google haven't done anything IMO.
They haven't even dealt with the obvious proxy scripts yet (nph,cgi-proxy,php proxy), which you can easily verify for yourself.
[edited by: tedster at 2:49 pm (utc) on Oct. 12, 2007]
1. This thread: Is Google fixing the proxy hijack problem?
2. Related topic: How can we defend against proxy hijacking? [webmasterworld.com]
I re-opened the linked thread where we can discuss our own defensive steps - and this thread can stay focused on Google's progress, or lack of it.
As I said above:
1. This thread: Is Google fixing the proxy hijack problem?
2. Related topic: How can we defend against proxy hijacking? [webmasterworld.com]
I'm not sure what the silence here about Google making any progress really means. Is the issue getting handled?