AlexK - 4:05 pm on Jun 10, 2005 (gmt 0) The language encoding can be established at many stages: by the web-server, or by the pages themselves. Perhaps the simplest is to put
I still can't believe that there isn't an easy fix to this: this simply cannot be based upon domain name and host location alone. Can it? There MUST be something else surely.
What is the language encoding on your pages? Is it en or en-gb or perhaps even utf8? It obviously wants to be EN-GB for UK-targeted pages.
<html lang='en-gb'> at the top of the pages, or even to use the META HTTP-EQUIV (which is intended to instruct the web-server to set a HTTP Header for the document):
<meta http-equiv="Content-Language" content="en-gb">The simplest way of all, of course, is to instruct the web-server to send the appropriate header for every page by default. A good tool to check the current headers for your web-site's pages is at Rex Swain's HTTP Viewer [rexswain.com]. If you find no Content-Language header then you immediately have something to fix.
You would use the robots.txt text-file to do that. This is a simple text file (normally located at the web-root) which is accessed by all legal-robots on all accesses to your site, and contains coded instructions on which parts of the site can and cannot be accessed by those robots. It can also be customised for specific robots. Look on the Google site for webmaster help for instructions - very simple.
we will probably try to exlude the site from Google altogether. Does anyone have the code(?) to do that?
The language encoding can be established at many stages: by the web-server, or by the pages themselves. Perhaps the simplest is to put