Welcome to WebmasterWorld Guest from 18.104.22.168
Forum Moderators: mademetop
Does anyone have any thoughts / opinions on how the eventual takeover of HTML by XML will affect SEO processes?
My understanding of XML is very minimal; I understand that custom tags can be specified for an individual site - this, I would imagine, will cause lots of trouble for spiders - anyone have any ideas how spiders will get round this? It certainly seems that they will have to become a LOT more advanced than currently if they are to keep doing the job they do.
Any thoughts / opinions would be appreciated.
joined:June 27, 2000
XSL/ XSLT allows you to 'pour' -- for lack of a better word -- your data with XML tags into an HTML template. I see it in my mind sort of like a CSS sheet for XML tags. You specify which XML tags go into an HTML template.
I've only played with it, but I'll try to fetch an expert. I am sure someone can explain it better than I can.
My suggestion is learn it!
I know this because I am SEO for a company who's site is built around ASP, XML and XSL. All of these transform easily (to the spiders and browsers) into HTML. The code is transformed.
XML + XSL = HTML.
Yup. The reason XML will get popular is that XML can also be used to exchange other data between apps. The XML>>HTML thing is just one example of what can be done. Having the ability to push/pull info from the web and your back-office systems lets you do some really quite neat things
In many server side case scenarios this is absolutely true. However, everything depends on your implementation. This would be true when using a XML app to compile server-side and deliver html (browser specific). Many off the shelf products do this now. You can define your own XML doc that is closer to what I mention above- pure text, very little code. The spider does *not* record HTML but XML. I have verified this independently.
As for resources, I like IBM, WDVL and CNET.
May I ask who these people who have verified this are? I work VERY closly with a number of the major search engines and directories and I can *assure* you that a spider will read the code that is sent to the browser NOT the XML (unless you fail to transform it).
I did not mean to imply they don't record what the browser sees. Send XML to a browser with Gecko or MSXML.exe and you're serving XML. That is all I meant. I apologize and didn't mean to confuse.
Since XML does not work without a parser, we're forced to use XHTML at the moment which is fully backwards compatible (in the W3C standards sense anyway). It works wonderfully and search engines record it (more) easily. My guess is XML will still transform the Web - it is only a matter of time.
What I'm talking about is the ultimate goal, down the road - when many web sites use XML (that will be translated via XSL into HTML, of course) search engines will be able to read and understand the XML of the page, not the HTML. And this will give search engines the ability to understand what you have writen, instead of just the way you have formatted it.
When XML is ubiquitous, search engines will answer your questions in the results page rather than linking to another page that answers your question - that's a fundamental difference. And this will only happen, of course, if TBL's vision for the web is actually realized. And that doesn't look likely at present - look at the abysmal rate of CSS adoption.