Forum Moderators: open

Message Too Old, No Replies

I don't know if anyone cares...

         

littleman

9:14 pm on Aug 21, 2001 (gmt 0)



...but NL is cloak sniffing.
HTTP_USER_AGENT = libwww-perl/5.50
REMOTE_ADDR = 205.181.75.60
Name: tantalum.northernlight.com

toolman

9:19 pm on Aug 21, 2001 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Oh No!
Oh My Gosh!
I've got to go....

littleman

9:23 pm on Aug 21, 2001 (gmt 0)



You know, at one time I had 60+ thousand pages in NL, and it was worth maybe 50 refs a day.

toolman

9:48 pm on Aug 21, 2001 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



>>>>60+ thousand pages in NL

So that means you had about 1/4 of their database and 1/4 of their daily traffic.

But why, oh why did you ever give it all up LM?

awoyo

4:44 am on Aug 22, 2001 (gmt 0)

10+ Year Member



Does this mean Fast is doing the same?

fpcr01.bos2.fastsearch.net (209.202.148.100)
This visitor used libwww-perl/5.53 FP/2.1.

littleman

5:50 pm on Aug 22, 2001 (gmt 0)



Could very well be.

idiotgirl

7:09 am on Aug 24, 2001 (gmt 0)

10+ Year Member Top Contributors Of The Month



So, does this mean that NL and Fast is looking to see if the pages are cloaked, correct? Which would include possibly parsing for SSI or embedded exec commands?

Idiotgirl

awoyo

6:08 pm on Aug 24, 2001 (gmt 0)

10+ Year Member



So, does this mean that NL and Fast is looking to see if the
pages are cloaked, correct? Which would include possibly
parsing for SSI or embedded exec commands?

I wouldn't think so. So many sites use SSI or embedded exec commands in dynamic situations that restricting their index like this would exclude perhaps thousands of sites that are not cloaking. They may be checking file sizes of pages that may be cloaked by User Agent, it may be a human looking at "suspect" pages, or any number of things.

WebGuerrilla

7:21 pm on Aug 24, 2001 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member




I don't know of any way a spider could possibly detect SSI. The command is executed server side before the page is ever delivered. Neither of the IP's are new, so the best they could do with it is catch UserAgent stuff.

Did either of you notice whether these bots requested a robots file?