Forum Moderators: Robert Charlton & goodroi
My pages are still listed as url only, so does anyone know what the deep crawls indicate, and when title/desc may appear in serps?
Thanks!
did you use the re-inclusion form that GoogleGuy said to use: [google.com...]
From your earlier posts I think I submitted a reinclusion request on the same day as you and started getting deep crawled 1 day after the reinclusion request (previously only 1 or 2 pages had been crawled for 30 previous) this was on the 22nd then 23rd and so on (last was on 30th).
However no title and descriptions have come back for that site and there are still many pages missing even though they have been crawled at least 3 times.
Another site of mine lost all title and descriptions around the same time as that site. The index came back after been re-written (2 weeks later) but then dissapeared again after the re-inclusion request which is a bit of a suprise as some titles and descriptions of sub pages came back 2 days after that reinclusion request and ranked (not as well as before but there all the same). Further happenings (after this) included my home page showing number 10 (and number 5 on 1 datacenter) for the sites name (which is keyword-1-2-widget) but showed URL only. The index is still missing for that site but have tried re-writing 1 or 2 paragraphes again.
I did have a similar experience a month earlier with the first site I mentioned, the rankings started to decline 3 weeks bafore a major update for no reason, so I changed the content drastically on the home page (it had been stale for ages) Then, after the next update my rankings shot back to above my previous position, I took it at the time that the update cleaned out the problem and the changes had been beneficial.
This time around I didn't make changes until 1 week later when the site had vanished to URL only. (too late)
From what I can sum up from my experience, and what others have said, is, there seems to be no order. So I don't think anyone can give a set time as to when and where Rome will rise again.
I was thinking that maybe a major update will be more helpful, but who knows.
So, is this behavior normal? Did my reinclusion request get accepted? My site remains indexed with internal pages showing url only, and index page listed with title/desc last cached April 22.
Chris
Gbot has not crawled my index pages of my 5 interlinked sites from last 4 to 5 months..Though some of my internal pages get crawled after every 10-12 days from last 50 days.
What does this mean?
How can I crawl my index pages...as it is very very important for me.
It is quite possible that Gbot has always "deep crawled" my site prior to the next update ...
Is Gbot just particularly active right now? Getting crawled again today. :)
Shortly before *every* major + minor update (at least since Dec 04), G has deep-crawled my site. Then, 24-48 hours later--oh dear--the update became active and the site's position in the G SERPS fell like a stone, as did the hits.
I can trace the recent rise in bot-hits back to about 9 Apr (only an estimate, as my site is on a temp server) but then *really* ramped up on Fri 29 and, Tue/Wed, is of the order of 5,000 pages each day. Blimey.
I can only hope that--this time--my no-tricks site will rise in the SERPS.
Slurp indexed them all...a day after letting the spiders in. I was shocked
The first (test) report is only for 4+ days (1-6 May), but Search bots account for 30% of the hits and 23% of the bandwidth:
Type --- Pages - Hits --- Bandwidth
people - 32907 - 160301 - 801.41 MB
bots --- 13955 - 14477 -- 235.86 MB
As only 36.8% of the site hits originate from a search-engine, I'm beginning to have my first doubts about their value.
What would happen if 36.8% of your traffic dryed up over night?
Further how many of 53.2% of your visitors started visiting because of a search engine?
We found that a large percentage of paying customers are short termers from search engines looking for a quick answer to a particular set of requirements.
The long term repeat visitors while they provide less revenue provide the required traffic to keep advertisers happy. Most of these visitors started as search engine referalls.
But if you wish a simple robots.txt file will cure this issue and we'll happily welcome your former search engine refered visitors.
Just let us know which subject area you are in so we can build some content;).
crobb305,
I noticed that Google has two of your pages showing title and description in its listing. I really can't remember what it was when I first looked. I thought it was only the home page then.
I've looked at couple of dozen sites in a number of niches, there was plenty of damage and there are still folks who haven't figured out what hit them.
It is just the size of the html document.
if I save my view source it's close to 40 KB. if I save (on IE explorer) as text, it's a little over 10. I wonder G doesn't take
What does Google show as the Cached size?
& is all of the page displayed in the Cache?
why would they pull the site 4 times? Gathering the links? Not that I mind :), just curios..
What would happen if 36.8% of your traffic dryed up over night?
... there was plenty of damage and there are still folks who haven't figured out what hit them....and you have? Do let us know.
walkman:
Google got my index 4 times last night and I'm wondering
crobb305:
My pages are still listed as url only, so does anyone know what the deep crawls indicate, and when title/desc may appear in serps?
The fundamental question of this thread remains: "How long til titles/descriptions appear?"
what does the number in bold mean:
HTTP/1.1" 200 8341
HTTP/1.1" 200 8093
HTTP/1.1" 200 8102
HTTP/1.1" 200 8037
if I save my view source it's close to 40 KB.That is possible if the page is compressed by the server (gzip/deflate). This page [leknor.com] will let you know for any specific page on your site whether it is gzip-ped (shows the Content-Length, too), and will show the headers for deflated pages.
why would they pull the site 4 times?Probably because it has varied in size every single time. I assume that you have dynamic content on the page?
What would happen if 36.8% of your traffic dryed up over night?
Further how many of 53.2% of your visitors started visiting because of a search engine?
My site started on a free-web-host many years back, and who cared who or how much bandwidth, since I did not have to pay the costs. Those days are long gone... Today, I have to pay for each byte downloaded.
With the search-engines it is a simple business proposition: they are allowed to roam at will all over my site because the cost to me of them doing this is offset by the return in visitors from their SERPS. When I then discover that 23% of my bandwidth is due to search-bots it stops me dead in my tracks. When I also discover that 83% of their SERPS for my site are url-only I begin to worry. When I discover that there is no means to find out why or to fix this situation I get really angry.
At this instant the SEs are a benefit for my site, but the equation is sliding rapidly towards the red side of the balance sheet.
I am gzipped ... with ob_ something ... wierd part is that my other sites are too but content length is different
will this mean anything bad to Google?My site is 87% url-only, but who the hell knows?
Some of the G-Bots are http 1.0, and therefore will not accept compression (HTTP/1.1 only). My site also provides a link to switch compression off for all pages (for anybody having display problems). I suspect that this is causing a duplicate-penalty.
"Cache-Control no-store, no-cache, must-revalidate, post-check=0, pre-check=0Do not quote me on this (check here [salemioche.com] instead) but this will instruct browsers (and proxies?) to re-fetch the page on every visit, which backs up G bringing your index page 4 x times.
Pragma no-cache"
All my pages are linked clean, but the PHPSESSID are for users to save a product for that session...not a cookie fan :).
Once again, none of my links are with /blah.php?ID=#*$!#*$!#*$!x, when I highlight or click I see clean links, and when slurp indexed the site, the links they got were clean (just /blah.php ).
Will this still be a problem for Google?
thanks guys,
How can I crawl my index pages...as it is very very important for me.
makes sense. Well my index, for good or for bad, changes everytime it's loaded...
This object has been deliberately marked stale. It doesn't have a validator present. It will be revalidated on every hit, because it has a Cache-Control: no-cache header. It won't be cached at all, because it has a Cache-Control: no-store header. Because of the must-revalidate header, all caches will strictly adhere to any freshness information you set. This object requests that a Cookie be set; this makes it and other pages affected automatically stale; clients must check them upon every request. It doesn't have a Content-Length header present, so it can't be used in a HTTP/1.0 persistent connection.
------------
The last modified line is blank. What else can I do? Google stopped by and picked about new 100 pages today, including the home page (full size too).