Forum Moderators: Robert Charlton & goodroi
Anything with Javascript is poor for SEO.
AJAX is also good to clean your bathtub, I'm sure there will be a few people who agree with me on this.It's also a brand name and I wonder how they feel about this.
AJAX makes really clean code - hehe
<edit>I seem to use to many ... now when writing...</edit>
[edited by: Pico_Train at 7:40 pm (utc) on Jan. 25, 2007]
We're all suppose to develop sites and content with the visitor's experience in mind, but I get cold feet doing anything with Ajax that might get confused as spam attempts, such as show hide panels with content within them.
Is gaceful degradation, where non js clients get delivered all the content that would otherwise be hidden an SEO friendly option?
So with the exception of forms, I don't have any experience with delivery user content through ajax.
This sampling appears organized as the bot reads exactly around 1.5 Mbytes of audio/mpeg content, then exits, waits around 15 secs. then reads again -- it repeats this as many times as necessary until it reaches the content-length of the enclosure.
This seemed to start once I added the feed to my personalized Google page -- this caused the bot to refresh the feed around every three or so hours -- the feed has a <TTL> tag of 10 mins. so this is ignored by the googlebot.
I guess that means that the googlebot can read and understand XML -- what it does or is doing with the content is probably up in the air now.
* I use JavaScript to hide stuff from search engine bots -- it's especially useful for affiliate stuff that is specifically intended for site visitors, and negatively seen by search logic.
We're thinking about hosting multiple topic-organized RSS feeds on an html page. The feeds will be pulled with JavaScript and the only real reference to the content will be in the html via .js.
My concern is that Google will not see anything on this page. It will not be googleable at all.
It sounds like this is the case. But many sites (including Google homepage) are creating content similarly. Will Google be able to evolve? Or is there a workaround on our end that will make the content spiderable?