Forum Moderators: open
Since LWP::Simple a module, whatever it's doing is up to whoever wrote the application it's been plugged into. It might be used for mailicious purposes, such as feeding a spam database, and it might not.
The is a method in LWP to change the UA name to whatever you want. If you're seeing LWP::Simple in your logs, then the coder hasn't bothered to do so.
I've been getting a lot of errors recently for pages like:
blog.mysite.com/archives/000256.htm%20/
with the space and trailing URL mysteriously added on.
I've scoured my site to try to find errant links, but have been unable to do so.
The user agent with all these errors is: lwp::Simple/5.69
So does this mean I shouldn't be worried about the 404s? Or should I still try to track down what's wrong with my links?
No legetimate search engine is using LWP::Simple as the user agent for their spider.
But "legitimate humans" can.
I wouldn't block that user agent. If someone uses Perl for "bad" purposes, he would probably also change the user agent to something else.
If someone uses Perl for "bad" purposes, he would probably also change the user agent to something else.
rduke
Welcome to Webmaster World.
I must confess this is a new criteria for NOT denying a UA ;)
The theme in this forum is identifying spiders. Most folks are aware that Ip's, referrers and UA's can be modified, omitted and even faked and yet spiders are identified.
The suggestion and assistance (off-topic) is shared to assist others in denying these non-compliant spiders.
Just the other day I had a visitor which downloaded a zip from a page they had not accessed. Although I didn't take any action at that time, I did make note of the IP and UA for future reference. Even though I only a have a few ZIPs for downloading, I may add zip to my deep-linking rules.