homepage Welcome to WebmasterWorld Guest from 50.19.206.49
register, free tools, login, search, pro membership, help, library, announcements, recent posts, open posts,
Become a Pro Member
Home / Forums Index / Google / Google SEO News and Discussion
Forum Library, Charter, Moderators: Robert Charlton & aakk9999 & brotherhood of lan & goodroi

Google SEO News and Discussion Forum

This 51 message thread spans 2 pages: 51 ( [1] 2 > >     
Google - latest on site: queries fix
Google Site: Query fix
Whitey




msg:711126
 11:28 am on Jun 3, 2006 (gmt 0)

Thought you's all like to be aware of these details on the latest fix with the site: queries.

An update on the site: operator
6/02/2006 07:28:00 PM

Posted by Vanessa Fox, Google Engineering

We've fixed the issue with site: queries for domains with punctuation in them. We are still working on site: operator queries for domains that include a trailing slash at the end (such as site:www.example.com/ ), so you may get better results for now by omitting the trailing slash in your queries. The Index Stats page of Google Sitemaps no longer uses the trailing slash for its queries, so you should see correct results when using this page.

Thanks for your feedback and patience.

[sitemaps.blogspot.com...]


 

trinorthlighting




msg:711127
 9:48 pm on Jun 6, 2006 (gmt 0)

I firmly believe that this issue is what is causing pages to drop.

nippi




msg:711128
 10:32 pm on Jun 6, 2006 (gmt 0)

except that the majority of people who have lost pages, don't use google sitemaps

LifeinAsia




msg:711129
 10:36 pm on Jun 6, 2006 (gmt 0)

except that the majority of people who have lost pages, don't use google sitemaps

And don't forget about all the people who don't have punctuation in their domains.

arubicus




msg:711130
 10:38 pm on Jun 6, 2006 (gmt 0)

Anyone who does use sitemaps getting this error:

Indexing summary:

This service is unavailable.

Please check back later.

AustrianOak




msg:711131
 10:41 pm on Jun 6, 2006 (gmt 0)

Yes I am getting the "Service unavailable" as well..

trinorthlighting




msg:711132
 10:43 pm on Jun 6, 2006 (gmt 0)

I meant I do not think the / is just effecting sitemaps only. I think its also effecting the way google indexes.

It would not suprise me if this is also a factor in dropped pages.

arubicus




msg:711133
 10:55 pm on Jun 6, 2006 (gmt 0)

"I meant I do not think the / is just effecting sitemaps only. I think its also effecting the way google indexes."

I am currious on that also. For one...why has it been taking this darned long to fix?

g1smd




msg:711134
 12:07 am on Jun 7, 2006 (gmt 0)

Search for an email address with a hyphen in it.

See that you get a bunch of results.

Replace the hyphen with a space, and search again.

See that a large number of supplemental results appear, and the snippet does show that same email address again.

More work to do.

dmje




msg:711135
 12:35 am on Jun 7, 2006 (gmt 0)

Is anyone else having trouble with the links on the sitemaps index stats page? I click on the link, a new blank window opens and nothing shows up.

I checked our site by copying and pasting the link from the sitemaps page into our default google (64.233.161.104)and show 770 pages without the trailing slash, 779 with the slash. Isn't this just the opposite of what Vanessa said would happen?

trinorthlighting




msg:711136
 2:42 am on Jun 7, 2006 (gmt 0)

It is probally taking a long time to fix, because it is effecting their whole system

whatcartridge




msg:711137
 7:12 am on Jun 7, 2006 (gmt 0)

I quite like the trailing slash thingy.

With trailing slash - shows pages on our site indexed by Google

Without trailing slash - as above but also includes supplementals

Whitey




msg:711138
 12:04 pm on Jun 7, 2006 (gmt 0)

The good thing from Google's point of view is that it's getting harder to distinguish what they mean to do versus what is actually not working.

To put it politely, i think nobody knows what's going on in certain areas of functionality, at Google or even out here on WebmasterWorld at times, which is a worry for trying to establish some stability - it could take ages.

I mean, look at this for example:

i do a site: query on our sitemap pages on one site and get 2 pages,

on another site with the same structure [ but unique content ] updated 2 months later i get a correct 28 pages. The first site is showing a drop in sitemap pages, but actually shows more pages cached across the site.

How can Google say that site map content which is used for assisting the bots through the site be excluded from caching [ or maybe I've missed a new innovation!?!? ]

F_Rose




msg:711139
 2:01 pm on Jun 7, 2006 (gmt 0)

We currently own a hyphenated domain name..

We still see NO CHANGES..

With a slash and without a slash...

24 pages from our site listed..

496 pages supplemental results..

Is this considered a fix?

malachite




msg:711140
 2:34 pm on Jun 7, 2006 (gmt 0)

We still see NO CHANGES..

No changes on our sites either in terms of number of pages showing up in the index of a generic google.com site: search.

All pages seem to be indexed when doing a site: search on various specific data centres, complete with unique descriptions for each page.

One thing I have noticed when using the generic google.com search as opposed to a specific DC, is that all pages which do show up display only the (same) general site description rather than the individual page description.

Has anyone else noticed this, or is it just me? :(

F_Rose




msg:711141
 2:36 pm on Jun 7, 2006 (gmt 0)

"All pages seem to be indexed when doing a site: search on various specific data centres, complete with unique descriptions for each page. "

Could you give me the IP Address of the datacenter?

malachite




msg:711142
 3:10 pm on Jun 7, 2006 (gmt 0)

There's a good list of apparently updated DCs in msg #45 of this [webmasterworld.com] thread. I'm not so sure they're truly updated - some of my results are pretty old, but at least the pages are listed, albeit with some as supplementals.

I also tried this one 216.239.59.104.

F_Rose




msg:711143
 3:17 pm on Jun 7, 2006 (gmt 0)

All the same crap..

Only 24 pages indexed from our site..

Some datacenters include 496 supplemental results to it, some don't..

Supplemental results going back 01/26/05.

malachite




msg:711144
 3:26 pm on Jun 7, 2006 (gmt 0)

Some datacenters include n supplemental results to it, some don't..

True. And doesn't help one iota if Joe Public surfer is getting the datacentre which shows only n (very small number) of results.

Weird thing is, Googlebot's been all over my sites in the last few days like some 06/06/06 demon! I gave up trying to fathom Google's algos after Florida, and am just thankful we get most of our traffic from Yahoo and MSN.

F_Rose




msg:711145
 3:35 pm on Jun 7, 2006 (gmt 0)

Do you use Google Sitemap?

Does it show when Google has indexed your site?

Was it before the fix or after?

texasville




msg:711146
 4:39 pm on Jun 7, 2006 (gmt 0)

I just noticed a difference. On one site I manage, all but two pages have been supplementaled or deindexed. One is my index page. The other is a popup page. What is funny is the popup page is dup content. It is a list of laws and regulations copied from an authority site for regulating the "widgets" this site sells. It is only for the convenience of my visitors and all credit is given to the authority site.
Now here is what is different about that page from all others. It is the ONLY one that points to my homepage as [mysite.com...] . All of the rest of the pages in this site point to mysite.com/...so is this something?
The only 301 redirect on my site is from non-www to www.
Now I also have a site that all pages point to mysite/index.html and it is fully indexed except for the links page, which I named directory html.
Both have google sitemaps.
Opinions?

mibrahim




msg:711147
 6:23 pm on Jun 7, 2006 (gmt 0)

I noticed something else regarding the site command. inurl and allinurl return a different number of cached pages totally different from both of the site commands with and without trailing slashes.

So now, I get:
site:www.mysite.com -> 5 pages + lots of supplementals
site:www.mysite.com/ -> 4 pages + lots of supplementals
inurl:www.mysite.com -> 8 pages no supplementals
inurl:www.mysite.com/ -> 8 pages no supplementals

I don't have mysite.com except in urls from my website.

malachite




msg:711148
 6:48 pm on Jun 7, 2006 (gmt 0)

I've just seen a slight change for one of my sites when doing a site: search on google.com - ie taking pot luck which datacentre it uses rather than using a specified datacentre.

BUT, although there are now significantly more results, it's still showing the general site description for each result instead of the individual descriptions I'm seeing for each page if I use a specific datacentre.

No change so far on the other sites.

F-Rose - no, I don't use Google Sitemaps, but I did recently add my own sitemap component in an attempt to get the bot through the site. Given Googlebot has been through the site like a dose of salts since I added it, maybe that's why more results are showing in the index. Hopefully this improvement will roll out across my other sites.

F_Rose




msg:711149
 7:29 pm on Jun 7, 2006 (gmt 0)

" I did recently add my own sitemap component"

Is it a regular site map which should be included on every site, or is it something else?

Could you be more specific about this?

malachite




msg:711150
 8:32 pm on Jun 7, 2006 (gmt 0)

Is it a regular site map which should be included on every site, or is it something else?

My sites run on Joomla, and I added a sitemap component made to run with it. No faffing around trying to design one myself, just uploaded and there it was done! I should perhaps have mentioned I did have a sitemap beforehand, but this one is much, much better.

Mainly did so for two reasons:

1. There was a debate going on speculating whether Google sitemaps were a good idea, and as I don't like giving G too much information, I preferred implementing the Joomla one,

2. There was another thread discussing disappearing pages being from predominantly level 3 and 4, so added site map so everything is spiderable at level 2.

F_Rose




msg:711151
 8:43 pm on Jun 7, 2006 (gmt 0)

Could you please send me a sticky mail with a link to your site map?

It would be greatly appreciated..

arubicus




msg:711152
 9:09 pm on Jun 7, 2006 (gmt 0)

"2. There was another thread discussing disappearing pages being from predominantly level 3 and 4, so added site map so everything is spiderable at level 2. "

I was apart of that discussion. Some webmasters who have created rather large site maps seem to have lost pages again. Since our site is much smaller than theirs (a couple thousand pages) we were able to split up the site maps for each of our 15 sections. The largest site map has between 100 and 200 links and seem to have no problem (knock on wood). The bulk of our site did get re-indexed and is now being crawled in full about every day. Rankings are slowly going back up. When our site does come back we may take those site maps and redo them so that there are no more than 100 links by drilling doen a level but for now they seem to be working to get our pages crawled frequently and indexed.

We do have a google xml site map that has been in place since it started. We never really seen a huge benefit crawl wise from it - never tried a plain text sitemap. New pages seem to be picked up a bit quicker without having to go through the site but that is about it. The onsite site maps seem to work alot better to get heavier crawling/re-crawling of existing pages. This is my take on it.

We were fortunate enough that google was typically crawling 3 levels. A 2 level site may have some problems since the site map would reside on level 2 and any links would be level 3 which Gbot may not be so quick to crawl but would not hurt to try (google has always recommended one anyway). Our site maps make sure that all links are on level 3 (again that will change at a later date) with a simple outline structure and no design elements.

F_Rose




msg:711153
 9:21 pm on Jun 7, 2006 (gmt 0)

"Since our site is much smaller than theirs (a couple thousand pages) we were able to split up the site maps for each of our 15 sections."

Are you talking about Google site map, or your own site map?

arubicus




msg:711154
 9:28 pm on Jun 7, 2006 (gmt 0)

In that statement I was talking about an onsite sitemap. We do use an xml sitemap submitted to Google sitemaps that has shown to have little effect. That sitemap incorperates ALL of the pages on our site. It isn't split up. I believe that Google sitemaps say up to 50,000(? - Can't remember for sure) links is ok.

The onsite site map has worked wonders so far (knock on wood again) and pages seem to be "sticking" now.

The discussion that was being referred to was using an on-site sitmap linked directly off the home page to where deep pages would reside on a higher level (level 3). For some it worked good and others not. Don't know the factors why though.

malachite




msg:711155
 9:57 pm on Jun 7, 2006 (gmt 0)

The discussion that was being referred to was using an on-site sitmap linked directly off the home page to where deep pages would reside on a higher level (level 3). For some it worked good and others not. Don't know the factors why though.

That's right, and that's what I did, except I linked the sitemap from the main menu so it's accessible from every page. The visitors seem to like it too ;) So with one click, they and bots can see a link to every article on the site.

This 51 message thread spans 2 pages: 51 ( [1] 2 > >
Global Options:
 top home search open messages active posts  
 

Home / Forums Index / Google / Google SEO News and Discussion
rss feed

All trademarks and copyrights held by respective owners. Member comments are owned by the poster.
Home ¦ Free Tools ¦ Terms of Service ¦ Privacy Policy ¦ Report Problem ¦ About ¦ Library ¦ Newsletter
WebmasterWorld is a Developer Shed Community owned by Jim Boykin.
© Webmaster World 1996-2014 all rights reserved