Forum Moderators: goodroi
The 'cached text' link will give a closer idea of how google 'sees' the page.
Bear in mind that Google doesn't actually 'see' a page, rather it interprets the HTML code used. Googlebot doesn't actually render pages at all to do this, so CSS is largely irrelevant.
Of course, the ability to view a page as closely as possible to a visitor's experience would be very useful to Google, so who's to say where they are up to on attempting this?
So is there a logical reason to use robots.txt to exclude Gbot that doesn't include the possibility of "customised" CSS files? ;-)
So is there a logical reason to use robots.txt to exclude Gbot that doesn't include the possibility of "customised" CSS files? ;-)
Well, the most appealing reason is that serving CSS (javascript, images etc) to spiders is wasted bandwidth.
Of course, there's always the possibility that the search engines might try to interpret CSS to check for spamming, in which case blocking access to such files could be deemed 'suspicious'.
My opinion is that CSS, javascript or indeed almost any externally linked files are not for spiders, although if you're a cautious type them I wouldn't exclude them.
So when gbot visits, it checks robots.txt; if it is excluded from example.css (or a directory containing a stylesheet) then it shouldn't retrieve the file.
When a searcher clicks on the cache of example.com, the base href tag instructs their client to prefix all relative links with http//www.example.com/ so the CSS would be called just as if they visited www.example.com directly.
So in this example, at no point does Google view the content of the css file, although for someone viewing the cache, their browser will download the css in order to render the page.