Forum Moderators: open
I have emailed Google's help email for re-inclusion and although the reply is a standard reply letter, it does state we do not have a penalty on our site.
I wouldn't expect Googlebot to index very deep if the highest PageRank on the site is 4. On the other hand, PR4 would normally be enough to get 200 pages indexed with static-looking URLs.
If the URLs are quite complex (eg. /index.pl?1=foo&2=bar&3=yin&4=yang ) then Google isn't likely to crawl so deep. If the URLs have a CGI parameter called id then I don't think they'll be followed.
1) What is the PR of your site?
2) How many backlinks do you have?
3) Any knowledge of having duplicate content?
4) What is the PR of pages that link to you?
5) Any back links from a penalised site?
6) Somebody dropping links to blogs - forums?
7) Java on site that may be putting Googlebot off?
8) Previous problems with the web Host or IP?
sticky me and i will send the URL.
All URL's static
Many sites seem to have this problem, the previous thread that was edited is at: [webmasterworld.com...]
One thing you might try doing is spidering the site yourself with Xenu, and doing spider simulator, robots.txt, & server header checks using the tools linked from here. Perhaps they will reveal some clue.
Check out your robots.txt manually, too, to be sure there's no weirdness there. One condition that results in URLs only being indexed (no titles & descriptions) is when a page is linked to but is put off limits by robots.txt. Google indexes the URL, but doesn't spider the page and hence has no more info.
<edit: Sorry, no email quotes please.>
Well now im lost what to do and now facing bankrupcy after five years on the net.
Hope google get what they deserve for tring to be "GOD" who do they think they are "GATES"
It was people and small companies like my self that got them where they are today, with us there would be no Google
[edited by: ciml at 12:18 pm (utc) on Jan. 28, 2004]
Previously, you were returning 403 for /robots.txt. Have you seen Googlebot requesting /robots.txt since it started returning 404 (and more recently 200)?
jdurkin, do you see Googlebot in your logs? I don't know if there's a good test for Google crawler connectivity to a server.
Perhaps someone has experience of this from the Godaddy <-> Google problem a while ago. Maybe Google translate or WAP gateway is a good test to see if the robot can reach a URL?
Tested me server for crawling -OK
I'm nacked chaps if something does not happen soon, just started paying server fees and a lot more bills out my own bank account, layed off two of my staff and google doesn't give a <snip>
No more sob stories from me chaps.
The worst thing of it all is pure frustration.
[edited by: ciml at 8:46 pm (utc) on Jan. 28, 2004]
I took the site away from our crap hosting company last week and all of sudden Googlebot is everywhere and our pages are returning to the index.
Conclusion is that our particular problem was being caused by our host. I am thinking they must have blocked Googlebot but never bothered to tell us.
The lesson I have learned here is how important Hosting is.
Thanks to everyone at webmasterworld who took an interest in my nightmare.
Sorry Googleguy for some of my posts over the last couple of months, its was nothing personal, just pure frustration.
We help you, you just ignore us.
Goglebot had been hitting the IP which I stopped using almost an year back. After I left the old IP... I had several deep Crawl on new IP.
I wonder what made Googlebot go to old IP again.
Whatever,
several sites are screwed like this without any clue and Google is speechless on that.
[edited by: ciml at 1:00 pm (utc) on Feb. 9, 2004]
[edit reason] Please see StickyMail. [/edit]
So this is exactly what happened to you, change IP google spiders it and then for some unknown reason starts going back to the old IP. Surely this is a problem.
Google maybe you should add a little more weight to the actual domain and not the IP.
Having pages in with "supplemental result" is worth what? They don't show in searches and as in my case where several months old.
With all due respects, GG has been actively involved in the discussions about the multiple URL listings in the serps. He seemed very surprised that this was occurring.
He has already flagged the appropriate people at g and they have agreed that a problem occurred but will take a few days for them to be removed from the current serps.
[webmasterworld.com...]
This thread is about "Pages still in with PR but no Description or Title!" GG has never commented on this issue which has affected many sites and is still an on going problem.
My reference to GG and the same URL appearing on each page of the SERPS was ment to highlight that it is unfair that GG asks webmasters for help and does often comment on problems faced by webmasters.
As far as the same URL appearing on many pages I personnaly could not care less, but for Google they cannot have this. Might make it seem that Google could be having problems.
*A chance to test my flame retardant armor ready and preparing for any fire to come* :D
Here is the exact quote from GG...
"I talked to the person about identical urls showing up and they found the problem. It will take a binary push, but should be fixed pretty soon (I would guess early to middle of next week)."
Try reading this thread from the start, its only 2 pages so should not take you long. We have all seen the thread about duplicate URL's in the SERPS and don't need a link to find it or need to be shown the quote he made. Again this thread is about "Pages still in with PR but no Description or Title!"
I've done some detailed research on this and if you search for the mt-blacklist you will find a lot of sites (including mine) on this blacklist (added for blog comment spamming) that have plenty of links to the site (as shown in allthewebs backlink check), but problems in Google. Either no backinks, no PR, no cache on indexed pages or some of the above while other things (like PR) is fine.
When my site went down it started mid december 2003 (I'd just got the traffic to 7000 unique visitors a day). Prior to the problems the home page peaked at PR7 for one month, but dropped to PR6, then PR5 backlinks also decreased in Google, but nothing to be overly worried about.
The site had and stil has over 6000 backlinks to the home page in Alltheweb. In Google it dropped to 2 backlinks and then to 0 over a couple of days. Traffic halved over a few weeks as pages were removed from the index.
At one point there was every page indexed (2500+), it has dropped to as low as 200, but now shows 600 but lots have no cache (about 180 have a cache).
The PR remained stable at PR5 until January when it dropped to Pr1 (still is PR1) which it clearly shouldn't be as I know there are PR5+ links to it.
I understand SEO (I offer SEO services) and so know what does and doesn't work with google. Other than the links to the site nothing else could cause this.
So has anyone else who has seen their sites PR/backlinks drop etc.. have a lot of spam type links to their sites?
David
1. allintitle
2. allintext
3. allinurl
4. allinanchor
The only the Top10 resemble or equal to is allineverywhere :D
Not one site of the Top10 showed up in any of the advance operator.
Interesting enough 'all' top10 are 'heavily' supported by spam links and anchors coming from spammed blogs, spammed forums, free web pages, all kinds of free account, rampant interlinkings, multiple throw-away domains.
You would think that this tricks have been filtered a long time ago. One spammy site might have slip the filter/s but the entire Top10? I didn't bother checking the rest of the serp.
Top10 is enough to convince me that the algo is way-off target.
BTW, I've been refreshing it from time to time hoping that this is just a local data center problem but it's sticky.
What can I say, spam links is the future of Google?
Reminds me of then defunct Snap Directory.