Forum Moderators: Robert Charlton & goodroi
I am scared to submit my site to google.
I have a site with 10,000 pages in which 4500 pages have been indexed.
Now i was told to submit my site thru Google sitemap to get the rest of the pages indexed by Google...
On my observattions i see that everyday at least 10-20 pages gets indexed by Google....as my site gets crawled daily.
Now what i was thinking if i submit my site thru Google sitemaps ....will that hurt the current scenario of daily indexing of new pages.
Or should i go ahead and submit the sitemap..
This is very critical issue for me coz i don't want to get in to any trouble.
Hope to get some good advices ...
Regards,
KaMran - White Eagle
That's the place...Always HAPPENING..
I was wondering.....will this sitemap tool will be good enuf for a site whose most of the pages have become supplemental which shows the date from JAN 05
i have made lot of changes after that but G has not picked the new changes after JAN.
Regards,
KaMran - White Eagle
Provided you have links to those pages, the answer is yes. In your sitemap XML file you should update the date of your last modification per page when you release the new version.
I was wondering.....will this sitemap tool will be good enuf for a site whose most of the pages have become supplemental which shows the date from JAN 05i have made lot of changes after that but G has not picked the new changes after JAN.
make sure that you are not returning 404's to google.
Have you deleted any pages?
If you delete a page or have a bad link the serve will keep sending "404 page not found"
This won't cause a page to be removed from the index for a long time, first it will lose the title and description - then it will go supplemental and bring some pages with it - in other words part of your site will go supplemental.
If your entire site goes supplemental then make sure googlebot is not being blocked for some reason - first thing you will lose all titles and descriptions.
So within 14 hours of submitting a sitemap, it's now crawled 95% of my pages. Still hasn't picked up the last 5% 24 hours later, but I have to admit that's pretty spiffy turnaround.
Nothing's been cached yet so I imagine it's still chewing on what it's crawled. That will be the big test I guess since just having Googlebot come by doesn't mean much unless the pages make their way into the SERPs. I'll let you all know how it pans out for me.
Will it hurt ...?
Well, I hope, it will not!
I can even think of positive effects of cleaning their index, as they can view the content of the sitemap as an authoritative inventory of a site.
Hey, Google, now that you have collected my sitemap, can you, please, check and compare against your supplemental index and silently drop all those old zombie pages and zombified 301/302 URLs that you may still hold for my site for ages, but which infact are long gone? Getting rid of all that olddated stuff could add to the index quality.
What counts, is in the sitemap, just forget about anything else.
I hoped that Google would do that, but some old stuff sits still in the supplemental SERPs ...
Would be fine!
Thanks and regards,
R.
I was wondering if I could get rid of old junk out of the serps using sitemap. What if I create a sitemap of everthing that no longer exists but still sits in the google serps. Will google visit these old pages, find they don't exist and remove them?.
This can be done easily by submitting robots.txt to the removal tool
I had them crawl a bunch of old junk pages the other day for the heck of it. Seemed to work fairly well.
Still nothing reflected in the Google cache 2 days after the big crawl. All dated June 9th. Will let everyone know when I see more happening.
That is what we are going to find out. We got alot of old junk laying around. Not touching the removal tool yet. We want to see what gets worked out naturally. But that old junk that does not exist no longer has active had my autogenerator put it in the site map on the next update. The next time google hit the site map it went nuts on all of those old junk pages. We just did this so we still are waiting to see what happens. But they will crawl it for sure. If you are impatient you can attempt the removal tool.
BTW - The site map works great for new content. Picks it right up on the next map download. No longer have to put a link on a frequently crawled page or have to wait for the bot to find it.
Have a very large site that has undergone a big update (during April and early May). Revision of url structure caused page numbers under site: command to drop to all time low as expected.
Also tidied up www canonical issue and ensured 404 pages gave correct header etc.
Added a Google xml sitemap soon after it was reported here. Site is updated daily and I use Googles login to submit the sitemap each day.
Last 48 hours site: command has jumped by 120k pages. However, traffic remains low - no perceivable increase in traffic even tho Google reports so many more pages.
Is this common? Do we just wait a bit longer? I thought maybe Google has added pages but not included any internal backlinks into the equation?
Has anyone got any similar experience and how long was the wait?
Thanks
Just to recap, I have a small site (50 pages or so at the moment), it was past all of the heavy filters (formerly known as sandbox) and pulling in a tiny bit of traffic in from Google. It was starting to get the cold shoulder from Googlebot, giving me a hard time with the new content I've been posting so I thought I'd give sitemaps a try.
I submitted my sitemap, within 14 hours practically all of the site had been crawled. It's now three days post submission and almost all of the pages are cached... All but one.
My index page is not cached... At all... It's still listed if you search for the URL, but when you check to see the cached version of the page it says "did not match any documents".
I'm assuming it's some sort of glitch, but it certainly seems strange as I've never seen this behaviour before. I have to say though that so far, Google sitemap seems to be working for me.
The week you quote - is that based on experience or a hunch? If experience, was it for a large or small site?
I'm not sure I understand what you mean by propogate across the web. All the content is being held in Google's cache but is not producing traffic. Do you mean propogate across data centers?
I would imagine that once Google does start ranking (some of) these pages at the top of the rankings, we may see some bonus traffic thru scraper sites but until that happens, I can't see any propogation happening.