Forum Moderators: open
I couldn't find anything about this on their site or in Google.
It did not read robots.txt and it did take disallowed pages.
This surprised me because Yandex is one of the few search engines I allow to crawl my sites because I get a lot of traffic from them. In the past they have always read and respected robots.txt.
YaDirectBot/1.0
77.88.57.nnn
nastenka02d.yandex.ru
[edited by: incrediBILL at 6:41 pm (utc) on June 9, 2009]
[edit reason] Obscured IPs [/edit]
93.158.136.nnn - - [13/Apr/2009:11:00:37 -0400] "GET / HTTP/1.1" 301 5 "-" "Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1)"
93.158.156.nnn - - [27/Feb/2009:05:44:03 -0500] "GET / HTTP/1.1" 301 5 "-" "Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1; SV1; FunWebProducts)"
Anyways for me the UA alone is not a factor to block access to a spider. However the target audience is and unless is international with english as the primary language it gets blocked.