Forum Moderators: open
I've recently started submitting a site to search engines and doing the usual link exchanges etc...
Its been spidered by the googlebot 3 times in the last week, but each time it just searched for robots.html and HTTP 1.0, it didn't even spider my index.html.
Is this normal? will googlebot be back to spider my site? or has it found something so objectionable that it doesn't want anything to do with it.
the site is in part a portfolio for my Director/Flash work, and therefore has a lot of supporting Javascript, but there's nothing really complex on the index page, just a couple of rollovers.
Thanks for any help.
Matt