Forum Moderators: Robert Charlton & goodroi
Google victim of redirect too ;):
Search for "Google" and [desktop.google.com...] shows first. If you click, [desktop.google.com...] redirects to Google.com
[edited by: ciml at 4:35 pm (utc) on May 9, 2005]
www.google.com is second for that search, not buried as is normally the case. So, can we compare this to a 302 'hijack' situation? Perhaps we can...
* [desktop.google.com...] redirects to [google.com...]
* [desktop.google.com...] come top for a search of "google", which we would not expect considering its links.
* cache:https://desktop.google.com [google.com] returns the page that should be returned for cache:http://www.google.com [google.com]
* [google.com...] has the ODP title and description - as can be case with pages that have been put back after a 302 hijack
* cache:http://www.google.com [google.com] shows a crawl date of ten days ago - much older than we would expect as can be case with pages that re-enter after a 302 hijack
[desktop.google.com...] is not well linked, and [google.com...] is well linked.
Could www.google.com have been down briefly when Googlebot tried to fetch it? If so, this would help explain why a page that redirects to it is crawled.
As GoogleGuy pointed out here, many people with 'hijack'-like symptoms had some other problem as the root cause.
I my situation I think I will wait until the new update, if nothing has changed, then its time to by a new domain, its said it has to be that way.
Nope! They have to submit an e-mail here [google.com...] and wait for the "Everything is normal /there's nothing wrong" response :).
Google probably got penalized. Most of their links are with the same anchor.
I still dont get why some are still out of the serps even if they have removed all googles junk and a month ago they made a supplemental DB update, where I saw all supplemental results where gone also on other sites, but a few weeks later that was returned.
All this trouble started for real when they added all those "sites" to there serps, then we saw alot of URL only (which before only came for a short time then the page was respidered with description), 302 links listed as pages, other domains in a site:yourdomain.com search (that problem is now gone because google just hide those now), %20%domain%20.com listed in google (why), I could keep on, but serious what happens here, why is google making such a mess out of there rankings/serps, we also got all those google sponsored scraper sites.
Plus, this is a redirect link to Google.com.
"Ummm - Could this result may simply indicate that they now place a much more emphasis on "new links"? The desktop is a great product and probably has a huge number of incoming fresh links from high PR sites."
how about finding the cure for everyone? If it happened to google with a trillion backlinks, what can happen to a site with 40 backlinks?
Now we can't even find who is linking to us with 302 because Google hides it.
AFAIK, a fix was never implemented. I do not consider hiding the wrong URLs in the "site:" search to be a fix, and I don't think it's a good thing to do either. FWIW, new hijacks still occur, old hijacks are not restored to normal, nothing has been "repaired", and Google is by-and-large still unable to tell the difference between an URL and a page1.
I am extremely puzzled about what it might be that makes a solution that hard to implement. I don't even mean "the solution", just "any solution", as several have been suggested. And, the many highly educated people at Google can probably think of a few things as well. They've had a few years to come up with something already.
By now, I can only interpret this as lack of will to get it solved. I don't mean to be rude, only that this is the only logic explanation I can think of. Not only that, it does actually seem like a strong will to let it remain unsolved at all costs - it's that strange to me that nothing gets done here.
---
1) To base choice of URL on PR is not a solution, it is a problem. There is no choice, as there can be only one right URL regardless of PR. As in: "I will not give you the keys to my house even though all your friends say that it is your house"
To base choice of URL on PR is not a solution, it is a problem
I've never really understood the logic to it either but I just assume they have their reasons for handling things that way. In any event, I have found it helpful to try to analyze what programming logic they must be using to decide which site they pick as the winner and then modifying the losing page (or sometime the whole site needs changes) to try to outrank the redirect page.
Not for me right now, not even with 100 results enabled. google.com/search is second, not google.com
I don't think it is a lack of will. It is a lack of ability to fix the problem. Some of Google's issues are so obviously of no value -- Supplemental Results showing for pages that haven't existed for more than a year, even though the 404 pages have been seen as 404 dozens of times -- that it must be assumed serious problems exist that they don't know how to fix. This may not be one, it may be a question of will, but I believe they know they have problems they wish they could fix.
> "I don't think it is a lack of will. It is a lack > of ability to fix the problem"
I agree with this sentiment. If they "fixed" this, something else probably goes wrong in the way they have the algo setup. We have to buy existing domains and try to cheat to survive.
I don't mean to be rude, only that this is the only logic explanation I can think of.
If you consider the possibility that many pages are losing out to redirects because the sites they belong to have penalties, then there really is no significant problem, from Google's point of view anyway, for them to fix. This would logically explain why they aren't fixing the problem. Perhaps from their point of view they are merely downgrading sites that deserve to be downgraded and their current algo and and quality standards for this year are working as planned.
I'm sorry but, what penalty does Google.com have?
Repeating this search, which in April didn't even show a Google page in the results:
Google Googol mathematical followed zeros coined Milton Sirotta
Now I get 338 results with a Google page at about position 50. As before the results above are on topic, but, shallow.
Message boards, blogs, I get a foreign page with the only English part the sentence about Google.
(I guess thats caused by my regional weighting).
An astronomy site with a clone of Wikipedia. A site about buying link exchanges with a page on Google, a site on their stock launch ....answers.com, Wikipedia page itself... a slashdot article.. another wikipedia clone, Arabic language blog, law site, edu site, German language page, SEO site, blog, air conditioner site,.....
eventually we get down to the one site which has good deep results for : Google itself!
I wouldn't mind if the pages above were full of information about "Googol" or "Milton Sirotta", but blogs, even foreign language ones?! Wikipedia clones that rank above Wikipedia?! Minor filler pages on SEO'd sites outrank Googles own page?
All very weird, the hand tweaked 2 & 3 word searches are fine but the lengthy searches are all very shallow results.
He put up a new page called (tadaa!) New Pages.
Ostensibly, this is just a list of 'new links' that he has found.
Its really just the same 1800 or so pages he 302-jacked previously, maybe a few new ones.
I don't know if it works or not, but here's his method:
Instead of full jacked URLs like [jacker.php...] he has
relative links like /sites/site#123.
A simple BASE=jacker statment or the like will complete each URL to the actual
jacking page, which is just a 302 redirect header.
This way, people can't nuke them like the old ones.
Stolen credit for content, PR and what-have-you would accrue to the 'new-pages' page
(assuming this works at all) and from there via regular links to his main pages.
I have to commend the jacker for his tenacity and creativity. There's just one downside
for jacker: Now all the pages he's been jacking are on one URL for easy reference.
Feel free to sticky me for that. - Larry
Kanoodle 3
Teoma 1
search.msn.com 1
Yahoo/Altavista/Alltheweb 4
Wisenut max 7 words (Google Googol mathematical Milton Sirotta brings it up at 1)
Gigablast 2
bbc.co.uk 3
ncgimaker 1 (my opinion)
They don't always give the same Google page, its either a press release or similar corporate information page.
By now, I can only interpret this as lack of will to get it solved. I don't mean to be rude, only that this is the only logic explanation I can think of. Not only that, it does actually seem like a strong will to let it remain unsolved at all costs - it's that strange to me that nothing gets done here.
I am with claus on this one. To start with it looked as if Google has fixed the 302 problem, then it slowly dawns on us that it is still there but buried/hidden, in as much as it is well neigh impossible to see if someone has put the hex on you or not.
No doubt some engineer, or even PR spin doctor, thought that was a great wheeze. Come on Google do something about it, or at least restore our means of finding the b*****s
I personally don't think all the intelligent people at Google can't come up with a fix to a simple problem like this. I have thought about it, but i have also ruled it out as that would make those people utterly incompetent, which is not my general impression. A very simple solution was implemented by Yahoo! some time ago, that wasn't impossible at all.
It may well be that it is not easy, but it is certainly possible. My reasoning goes like this: If you've got a problem that can be solved, and you don't do something that can be done to solve it, it must be because you don't wish to solve it.
And that's what puzzles me, because there's no apparent reason that these errors and malfunctions should be desirable.
>> If you consider the possibility that many pages are losing out to redirects
>> because the sites they belong to have penalties
You can also choose to view this statement that GoogleGuy made in another way: That the 302 hijacks cause a "pollution" of the target domain, which is why they get "penalties" in the first place.
So, they are getting penalties because of the 302 redirects, not the other way round. This also explains why sites don't re-surface after hijacks have been removed.
>> q=google
I also saw the listing as originally reported yesterday. It has been fixed now.
So, they are getting penalties because of the 302 redirects, not the other way round. This also explains why sites don't re-surface after hijacks have been removed.
Like maybe a duplicate content issue?
Now if Google has exactly one way of looking to see if a site is really what it is, then fixing the site: view should also start the timers on removing the duplicate content issue.
How long is a page "filtered" for duplicate content issues?.
Is by number of times (how do they count how many times), length of time, number of copies of, or some involved combination thereof?
So, they are getting penalties because of the 302 redirects, not the other way round. This also explains why sites don't re-surface after hijacks have been removed.
When sites don't reappear after the 302s have been removed, then it would be logical to consider the possibility that the 302s were not the root cause of the problem in the first place.
"When sites don't reappear after the 302s have been removed, then it would be logical to consider the possibility that the 302s were not the root cause of the problem in the first place. "
When sites don't reappear after the 302s have been removed, then it would be logical to consider the possibility that the 302s were not the root cause of the problem in the first place.
Assumpions when dealing with large systems are usually wrong. Note the use of the weasel word "usually".
What may appear to you as the 302 having been removed may not in fact be the case.
or in the alternative:
Once a page has been declared a duplicate it will take time for that designation to wear off.
If the clocks just started and the declaration was made 6 months ago you could be in for a long road back.
YMMV and logic has no bearing on the matter if you can't see the actual code that is functioning, everything is an assumption.
I've been waiting for years for 'corporate media' to take a stab at taking over the web and bouncing all of us small time publishers.
So suppose that the greatest and most respected SE, one that seeks to do 'no evil,' was bought out by corporate big media interests.
Furthur suppose that this SE gets more and more filled with spam; spam which this magnificent SE can't seem to stop. In a final effort to 'clean' the index and 'take back' the web from 'spammers' the SE introduces a new criteria called Trust Rank. In order to compete in the world of TR, a web site will need to be interlinked with big time 'trustworthy' corporate controlled sites.
Result: small publishers disappear forever, the searching public thanks the SE for fixing the 'spam' problem and the corporate take over of internet eyeballs is complete...