Forum Moderators: open
Deepak's example should be followed by more robot users.
Jim
http:// www.infoplease.com/
or
http:// en.wikipedia.org/wiki/Main_Page
?
As related to the content of my sites, all these two do is distract visitors with links to insufficient explanations, while waiting for somebody else to add plagarized data from my sites :( ALL the while somehow grabbing irrelavant PR's.
Even the "this day in time" type pages although enjoyable and perhaps beneficial, in most instances provide sparse data, while detracting visitors who are in most instances looking for depth.
JMHO
Here's a prime example which is nothing but keyword spamming:
http:// en.wikipedia.org/wiki/Hambletonian
The links go nowwhere in most instances while awaiting some unknowing soul to come along and copy data from another source in which credit is never provided.
or is deepak gathering data which will perhaps help SE's in the future provide some significant explantion of what the following means?
Knight Dream, p, 3, 1:59
:):)
[edited by: volatilegx at 5:55 pm (utc) on July 27, 2004]
[edit reason] delinked URLs [/edit]
I think he did a very good job on the user-agent string and on his info Web page - Lots better than many commercial robot users provide. Call it enlightened self-interest if you like, but at least he *is* enlightened regarding the fact that Webmasters don't like "mysterious visitors with inknown intent."
Jim
The old "keebler thread" in the archives is a prime example. When I initiated that thread I was unaware of the real source of all the prodding and poking around. I later in a land-line conversation realted to an article I used as an example in one of the threads, found out that all the prodding was the result of a MAJOR corporation seeking data for their 100-year anniversary. They besiged the same Public Library-Historical Center that I retrieved some data from with a couple of dozen techs gathering data or at least what they could find (path and keywords unknown.)
All these techs getting paid good money from said corp.
To think that all this nonsense, secrecy and expense was generated from an accidental reading of my page by somebody at Keebler and how much expense was wasted when all they needed to do was to send a cordial email inquiry :)
Had I not stopped their spidering of my sites, they would have likely still been crawling. All the time at the expense of my bandwidth and my research.
I do realize that my example is extreme as related to most websites. However, severity doesn't lessen what takes place or how actions should be reacted to, at least IMO.
As previously stated many times, each webmaster must determine what is both benefical and detrimemtal for their own website and visitors.
I personally see no difference between university research and other bots who collect and provide data to 3rd parties under a fee-based umbrella.
BTW, I did read some portions from a couple of his papers and found it quite interesting, in fact, one portion which I read provide nearly a duplicate example of the question I asked at the end of the previous post.