Forum Moderators: Robert Charlton & goodroi
Google Updates Webmaster Guidelines: Crawling Page Assets May Help SEO
As our crawling and indexing systems improved to render pages as part of normal indexing, today we're updating our webmaster guidelines to reflect that. Our new guidelines explain you should allow Googlebot to crawl the page assets (CSS, JavaScript, etc) so that we can index your content properly.
Let me be super clear about what this means: By blocking crawling of CSS and JS, you're actively harming the indexing of your pages. It's the easiest SEO you can do today. And don't forget your mobile site either!Google Updates Webmaster Guidelines: Crawling Page Assets [plus.google.com]
We recently announced that our indexing system has been rendering web pages more like a typical modern browser, with CSS and JavaScript turned on. Today, we're updating one of our technical Webmaster Guidelines in light of this announcement.
For optimal rendering and indexing, our new guideline specifies that you should allow Googlebot access to the JavaScript, CSS, and image files that your pages use. This provides you optimal rendering and indexing for your site. Disallowing crawling of Javascript or CSS files in your site’s robots.txt directly harms how well our algorithms render and index your content and can result in suboptimal rankings.Updating Google Technical Webmaster Guidelines [googlewebmastercentral.blogspot.com]
Disallowing crawling of Javascript or CSS files in your site’s robots.txt directly harms how well our algorithms render and index your content and can result in suboptimal rankings
Planet13 wrote:
Hopefully that means those sites with the annoying css-based pop-over ads that take up nearly the full size of the screen will start to loose rankings.
I don't think there is a valid argument against allowing googlebot access, as - if you don't - Google has no legitimate way of knowing whether what you are restricting access to is good or bad.
However, it looks to me that I will have to rethink my js navigation menu
Google has no legitimate way of knowing whether what you are restricting access to is good or bad
if Google is not indexing closed carousel / closed tabs (which can all be easily revealed on click, but do not generate a separate URL) doesn't this mean that they will suddenly lose quite bit of web content they currently have indexed?
Eric Enge: If someone did choose to do that (JavaScript encoded links or use an iFrame), would that be viewed as a spammy activity or just potentially a waste of their time?
Matt Cutts: ...In my experience, we typically want our bots to be seen on the same pages and basically traveling in the same direction as search engine users. I could imagine down the road if iFrames or weird JavaScript got to be so pervasive that it would affect the search quality experience, we might make changes on how PageRank would flow through those types of links.
It's not that we think of them as spammy necessarily, so much as we want the links and the pages that search engines find to be in the same neighborhood and of the same quality as the links and pages that users will find when they visit the site.
Ah, the old 'if you've nothing to hide, you've nothing to fear' argument.
I think Blend27's point about hacking is the best objection I have seen so far to the lack of counter-argument, but I'm not sure it works: hackers are not going to obey robots.txt anyway.
Allow-noindex: /scripts/
Is anyone aware of any online tools that can easily check whether Google Bot does have access to the "page asset files" that are discussed here?