Forum Moderators: Robert Charlton & goodroi
Noticeably, the interface is a little different. Sitemaps are now listed as "See stats for: yourdomain.com" Clicking this link shows several new stats (used to only show the crawl stats)
Now you are presented with this:
Query stats ¦ Crawl stats ¦ Page analysis ¦ Index stats
Here are the descriptions:
Query Stats:
Top search queries are the top queries to Google that return pages from your site. Top search query clicks are the top queries to Google that directed traffic to your site (based on the number of clicks to your pages in our search results).
Crawl Stats:
These statistics provide distribution information for pages we have crawled. [This page also shows page rank distribution across all pages crawled.]
Page Analysis:
These statistics show you how the Googlebot sees your site.
Index Stats:
You can use our Advanced Search page to find out how your site is indexed. Below, we've done some of these searches for you. [Just quick links to site:domain.com, etc.]
I like the information being provided on query stats the most, it's only a little glimpse, but lets you know what keyword you are showing up most for.
There is also a new "Errors" tab that has the following info:
HTTP errors ¦ Unreachable URLs ¦ URLs restricted by robots.txt ¦ URLs not followed ¦ URLs timed out
In order to use any of the advanced statistics, you must verify your domain. Information on stats and verifying can be found here [google.com]
The stylesheet is different, so it required a hard refresh when I visited the page. But other than that, well done. :)
What's new:
There are new "query stats" that show top Google search queries that return pages from a site. It also shows the top search queries that drove users to click on a site.
Crawl errors: Will now show, for example, specific HTTP errors Googlebot runs into when crawling a page.
Looks useful to me.
Log in here [google.com...]
"Data is not available at this time. Please check back later for statistics about your site."
especially the
"The PageRank of your pages in Google ¦ Distribution"
With me: the majority is "Low" ... #*$!IGO?
Nevertheless: thanks for the heads up, not everybody is drinking as it seems :-) ...
p!
If visitors reaching your page through a query string spend considerable time at your site, then google would regard it as an indication of relevancy, and would increase weighing factor of the string in question for your site. Conversely, if the visitors spend little or no time, then google would suspect that the association between the query string and your page is weak, that the relevancy is low, and accordingly will decrease its weigiht.
IMHO, this new feature is a real breakthrough in Google's fight against spam. It can easily turn out to be the spammers' nightmare if my assumption about the use of time parameter is correct.
As far as I can tell 'low' pagerank counts as either a 3 or a two. Not quite sure which though.
A shame it doesn't explicitly state which pages googlebot has hit though, I've never been able to figure that out except for a 15 day trial of analyse spider.
I signed up today and it’s pretty sweet. For example, you can now see crawl errors, timeouts on pages, robots.txt errors, unreachable urls, etc. Just really useful hard data that tells you if you have crawl problems and what they are. And you do not need a sitemap to use this functionality. You just create an empty file to verify that you own the domain.
And you do not need a sitemap to use this functionality. You just create an empty file to verify that you own the domain.
That is absolute gold-dust...
Questions:
Are all stats shown if you don't have an empty file (currently I only see query stats - but will that change?)
What's the difference between " Top search queries" and " Top search query clicks"... I'm seeing terms which my site doesn't even come up for under the latter column.
Problem with the system:
It mixes image and search click results (at least) but has a link only to the main search results. This makes isolating the two very difficult.
[edited by: vincevincevince at 10:44 am (utc) on Nov. 17, 2005]
... because it tells google who owns what domain (better than WHOIS data ever could)
That's a good point. Remembering the smart pricing approach, penalties could now easily be reflected across all your sites. A good case for multiple sitemap accounts?
I think I'll be creating one account per website unless Google can make a statement to the effect that they will not use the data we verify here to relate one site to another, and I'd suggest others do the same.
I'm getting this for most sites (no sitemap, but are verified):
Data is not available at this time. Please check back later for statistics about your site.
It's great! I like the "Top search queries" vs "Top search query clicks" feature. That's data that hasn't been available before.
Too right it's great! So what I'm assuming it's saying is "here's a list of the queries you're turning up in the strongest" (presumably a combination of position and frequency of the search?) and "here's a list of the queries where people are actually clicking on your result". Obviously we know the second one from our logs, but the first bit of information is fantastic. It shows one of my sites to be appearing strongly for the query "widgets" (and indeed, on checking, it does). But I never knew about this, because nobody ever clicked on my Google result. If I just go away and play with improving the presentation of that result, perhaps I can change all that!
If visitors reaching your page through a query string spend considerable time at your site, then google would regard it as an indication of relevancy, and would increase weighing factor of the string in question for your site. Conversely, if the visitors spend little or no time, then google would suspect that the association between the query string and your page is weak, that the relevancy is low, and accordingly will decrease its weigiht.
I don't think that would be the case as there is no way for google to know the amount of time that a user spends on any particular site. Unless of course that site is running the new google analytics code or using adsense.
It would be pretty foolish for google to add a weighting factor to their algorithm that they could only determine for a fraction of their index.
Really? Are you talking about what used to be the old well reputed search facility some university guys set up or are you talking about the exceptionally large business that seeks only one thing as all business entities do ... profits?
... because it tells google who owns what domain (better than WHOIS data ever could) and therefore who is cross-linking all their sites, so that they can take this into account when assigning values to links...
I disagree.
I have 12 domains in my sitemap account. My company owns 3 of them, the rest are our clients.
www.domain.com/site/
The sitemap statistics tool will not let me access query stats, crawl stats, or page analysis. The error message reads:
"If you verify at [domain.com...] we will add it to the Site Overview page and show you errors with URLs for the entire site, as well as a greater variety of site statistics "
But I don't want site statistics for [domain.com...] I want them for [domain.com...]
Google treats www.domain.com and www.domain.com/site/ differently in the index and I actually do not have access to the main corporate directory. Very frustrating. And I was so hopeful.
We haven't submitted a sitemap because we get crawled just fine. Does anyone know if the results are better if you do have a Google sitemap?
Is anyone experiencing this same problem? Is there a workaround?
I sent Google a note asking how to get my site stats, but I have little hope I will get more than an autoresponse.