Forum Moderators: open
The wierd thing is, I have a client's site that uses the same type of java scripting, and the Goo didn't make it past the first page the first time it got crawled.
baffled,
Chris
You should, however, provide alternative navigation for viewers with JS disabled. You can do this with a <noscript> tag. I use an iFrame in the <noscript> which embeds my sitemap to every page. This is seen by SE, and indexed as 1 outgoing link.
http: *//www.domain.com/cgi-bin/redir.pl?id=00012345
/claus
<edit>if, on the other hand, you want Gbot to follow your redirect, a spider-friendly redirect script can easily be made, there are other threads about this</edit>
Does Google see the text (when text is used instead of an image) in a Javascript navigation menu?
Does Google see the link in a Javascript navigation menu?
If so, when does Google not see the link? (I saw it written that absolute links are followed and relative paths are not, is this true?)
Can it follow this link?
Can someone give examples of crawlable web sites? (I hope this is okay to ask)
I appreciate the help!
The Java comment wasn't directed at you, it was a general reply to all those on the thread calling it Java, it just happened to be after your post.
I've stickied you the url for a site that uses JS links and has been fully crawled (no link posting on here).
Hope this helps, I don't know about the link text, I asked that same question this morning on another thread but nobody has replied. They seem to give the same PR benefit though.
Cheers
Nigel
I've seen some posts (not only from Made_In_Sheffield) suggesting that some links that are not on the standard form <a href=""></a> can get catched, eg. unanchored links or links in scripts, but i really do not know what Gbots exact limits are on this point, i have seen quite a few questions, but no really clear answers.
If you want it to be identified as a link, i would suggest you make the string as "link-looking" as possible and include the http part as well:
ie. "http://www.example.com/page.html", not just "page.html"
If, on the other hand, you want to hide it, it is quite easy to make JS-code that makes it hard to recognize it as a link. Just mess all the characters around and you're there.
/claus
I will add some definitions to let all of you understand what is Java and JavaScript:
Java: Object Oriented Programming (OOP) language based on C++ that lets the page designers to create small applications (called applets) that run on a web page independently of the platform. An applet must be compiled before it can be run.
JavaScript (or JS): Scripting Language for the Web that has a sintax inspired in Java, but very simplified. It is interpreted, not compiled. It's used, mainly, to make the content of a page more dinamic without cgi nor Java.
So, if I see again something that says Java to refer JS I'll be very angry X(
<head>
<meta name="robots" content="index,nofollow">
</head>
The problem with this approach is that the nasty bots couldn't care less what you write in a meta-tag. Not all bots are good bots. Gbot is, so the metatag would be better advice in this case.
It is a catch-all, though - the "nofollow" will be the rule for all links on the page, not all links except the secret link to my personal photos, notes, or whatever.
/claus
FWIW Matt cutts of google at Pubcon 4 said google is aware of Javascript links and can follow them.
I had an extended conversation at PubCon with Matt, and I can confirm this report. In fact, he said we can expect to see more of this as Google continues to develop this approach -- not parsing the javascript, but looking at it as pure text and seeing if their might be a URL in there.
Same thing applies to form actions and other nooks and crannies where potential URLs may hide.
Many websites do this for reasons other than hiding links. Personally, I like to test different text copy on my button labels, and using an external file allows me to do this in one place for a 100 page site.
So cloaking, cheating, whatever you want to call it is not always the intent of the website developer. Google will never treat this as cloaking for this reason.
At this years WebmasterWorld PubConf I asked Matt Cutts as to whether using Javascript for linking would be treated as hidden links. And he gave me a very clear no. He told me that using Javascript to link was fine.
He did make other comments about the text contained in an on-page javascript tag as being "interesting" and something that they were looking at further.
Make of that what you will.
I have found at least three threads here stating that Gbot does not index more than the first 100Kb of a file. It's a bit unclear if the first 100Kb are indexed for the cache only, or if the index have the same limit - the index-limit need not be as "low" as the cache-limit. Unfortunately, i could not find this information on Googles own pages.
Number of links:
There's apparently a limit of 100 links per page. The Google Guidelines [google.com] also mentions this, but not as a limit, just advice.
/claus
<added>Btw, thanks mil2k, Made In Sheffield, tedster, and egomaniac - it's really nice to get the JS-link-extraction confirmed from four sources :)</added>
Google follows text links very well and generally does not screw it up. Java seems to have mixed results in WebmasterWorld, so why would you need to do it. If you do need it for smooth opperation of your site use alt tags to make it less painful.
Please tell me that you are not using frames.