Forum Moderators: open
I've worked with lots of dynamically driven sites, but this one is a new one. Any insight on how Google treats these pages - and insight on how to get the dynamic content indexed, would be greatly apprecaited.
Your clients page would not typically look like this in the address bar:
[blah.com...]
If they're depending on the client (browser) to apply an XSL to existing XML data, then it would look like this: [blah.com...]
And then the XML file would reference the XSL file, which is used to render the final output HTML.
Or, more commonly, an XML DOM such as MSXML is used to apply an XSL to an XML document, and output HTML. In this case, the URL would look like this, and Google would index it as normal:
[blah.com...]
So is it server-side or client side XML processing?
They have several pages in the Google index, but less than 50, definitely not good saturation for a site of this size. I think they may have, in the past, been involved in some cloaking activity, but I haven't received confirmation of that yet.
It looks like Google is reading the pages, but only when .xsl is followed by database query language (? and =).
My guess is that Google won't read any of the main pages because of the .xsl extension, but will read some of the dynamically generated pages because of the query language.
Any thoughts on optimizing a site like this?
View the source of one of the pages, and let us know if it looks like this:
<?xml version="1.0" encoding="ISO-8859-1"?>
<xsl:stylesheet
version="1.0" xmlns:xsl="http://www.w3.org/1999/XSL/Transform"
xmlns:msxsl="urn:schemas-microsoft-com:xslt"
xmlns:user="urn:my-scripts"
exclude-result-prefixes="#default">
<xsl:output method="html" indent="yes" />
<!-- blah blah -->
<META http-equiv="Content-Type" content="text/html; charset=iso-8859-1">
But WHY? The general public could give a crap whether the page is written in HTML, ASP or Klingon for that matter, as long as it works. And these pages do work.
There could be several answers, but it's really strange. Why would someone do this?
But WHY? The general public could give a crap whether the page is written in HTML, ASP or Klingon for that matter, as long as it works. And these pages do work.There could be several answers, but it's really strange. Why would someone do this?
Sounds like he's reaching for some kind of dork-webmaster street cred, even though most would realize .xsl isn't what he thinks it is.
We may be getting off on a tangent here, but I don't know WHY they would do this
Bcoz search engines are the last thing they have on their mind. Also the desire to be different than those ubiquitous .htm files ;) Added to that ignorance is a bliss :)