Forum Moderators: open

Message Too Old, No Replies

Redesigned Site does not attract googlebot

Seeks index and robots.txt file and leaves

         

cyberceo

3:58 pm on Feb 15, 2004 (gmt 0)

10+ Year Member



I am a new entrant in this forumn and would like to thank everyone for their invaluable suggestions and postings. I've read Brett's "Successful site in 12 months" postings, have read googleguy's recent postings and found it all so very interesting.

I am seeking advice on the current problem I am facing. I recently redesigned (3 weeks ago) my website and changed the entire structure. Prior to this redesign effort, this website had not been touched since May 1997. There are several old pages from this website in the google index, which are now obsolete and generate 404 errors.

On my newly redesigned website, Googlebot visits it and leaves after fetching only the robots.txt and index file. It simply refuses to spider the entire website.

FYI, As suggested in the posts on this forumn, I frequently change my index page (adding content/links) and modify the robots.txt file in order to seek the interest of the googlebot. However, nothing seems to be working.

My site has a page rank of 4 and a few incomming links from PR 5-6. My site navigates well on lynx.

I'd like to abide by the rules of this forumn and not place my URL in this posting. If anyone wants to see the site then send me an email.

I want the googlebot to crawl my site and include the newly designed web pages. I also want my older files deleted fom the google index.

yowza

8:35 pm on Feb 15, 2004 (gmt 0)

10+ Year Member



I'd get some deep links for sure.

Also, adding any new links to the home page couldn't hurt either.

However, I would imagine that any site will eventually get reindexed if it is already in the index.

Send me a sticky with the url and I'll take a look at it.

kaled

11:30 pm on Feb 15, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Are pages dynamic (e.g. php) or static (e.g. html)?

Google doesn't care, but if dynamic, are you certain that all the headers are constructed correctly?

You said that you've changed your robots.txt file several times. Why not try using a blank robots.txt file. Also ensure that your existing one validates. (Be aware that a missing robots.txt file may result in an error pages being served instead.)

You should also validate the your index page. I have never used Lynx but my understanding is that it is a browser not a validator. There may be a badly formed tag somewhere that causes Googlebot to puke.

Kaled.

a_chameleon

5:53 am on Feb 16, 2004 (gmt 0)

10+ Year Member



Try [searchengineworld.com...]
and see how things look..

cyberceo

10:00 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



Thanks so much for everyone that sent me a sticky mail. I'd like to clarify that I have checked the robots.txt file using the sim spider as well as checked the apache configuration.

Further to add that all the currently indexed pages from my site at google when one goes to "site:www.sitename.com site" are giving 404 errors except the home page. The older files have been deleted.

Since we have created newer content and totally changed the way we run the business, I am unable to set 301 or 302 redirects.

Thanks once again to all who responded. My fingers are crossed. I hope the site gets indexed some day.

storevalley

10:13 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



Since we have created newer content and totally changed the way we run the business, I am unable to set 301 or 302 redirects.

One of my clients had a very similar problem a year or so ago. Two things that can help are ...

  • Redirect all 404s to your site map (you do have a site map I hope!) This will help spiders (and visitors) to find your new pages more easily.

  • Time ;) It took 6 or 7 months for Google to completely re-index the site. Some of the other search engines still haven't completed this yet, but they are getting there slowly.

thevoodoo

10:22 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



CyberCeo,

I would really think about repopulating my pages with the old URLS again, if waiting a few months is not an option.

Otherwise sit tight and wait a few months. Your new pages will eventualy be indexed and the old ones will drop. However since many of us have seen 1 or 2 year old removed pages come back into the Googl eindex after the past few updates, I still recommend somehow keeping the old file names.

One thing though: I am nto sure if setting all the 404's to the sitemap and letting it get indexed is a good option. YOu might need to make sure if you do that, theat it does nto get indexed, otherwise you would have a lot of duplicate or identical pages which is not a good thing! ;)

TheVoodoo

antsaint

10:29 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



Agreed - killing off those old pages can't be helping, since they're already indexed. Is there any way you can breathe some new life into them - at the least a redirect, but maybe find some way to incorporate them and their content in the new site? Perhaps just some gateways, fluff content, whatever.

storevalley

10:36 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



YOu might need to make sure if you do that, theat it does nto get indexed, otherwise you would have a lot of duplicate or identical pages which is not a good thing!

Seems a little unlikely that you would be penalised for using a standard mechanism for dealing with non-existent pages ...

thevoodoo

10:43 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



Would love to agree with you StoreValley, but taking the latest updates and the current SERPs including the ones on the 64.* DC, I must say you can not be too google-cautious or paranoid when it comes to dealing with pages and content on your website.

AntSaint, Those are even more horrid than using a unified Sitemap. Those methods suggest depricate spamming methods.

CyberCeo, I think if you email google with your problem and wait a few weeks you wil receive a response. and then you can enlighten all of us if the sitemap method would be harmless according to google. You will be amazed at how much specific help Google will provide if you email them patiently.

storevalley

10:52 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



I must say you can not be too google-cautious or paranoid when it comes to dealing with pages and content on your website.

Updates come and go. I just try to forget the paranoia and build web sites that are interesting and on topic. Works for me ;)

thevoodoo

11:36 pm on Feb 16, 2004 (gmt 0)

10+ Year Member



Storevalley,

That is the best practice. However keep caution lingering in the back at all times. Because even if you are innocent and not trying to do anytyhign wrong you can get slammed by updates. That has happened to my personal site with not even a single trace of SEO in it.

Unfortunately the internet, much like the world we live in, is full of spammers and crooks and it is always imperative that good people get burnt along side the bad ones. Not all of them get burnt ofcouse, but you always want to be wearing your anti-flammatory gear. hehe

Thevoodoo

storevalley

8:39 am on Feb 17, 2004 (gmt 0)

10+ Year Member



Because even if you are innocent and not trying to do anytyhign wrong you can get slammed by updates.

Quite right, thevoodoo.

So what's the point in worrying about it? Building a clean, on topic site without using any tricks means that you are creating the type of content that the perfect search engine would love.

And all of the search engines are striving to be perfect ... right? ;)