Forum Moderators: open
Site validation pending..
Yes, I agree that the idea of an authenticated dialog with site owners does sound a lot like the Google Sitemaps Program -- now renamed as Google Webmaster Tools. But I doubt this is a case of copying. In my opinion SiteExplorer was already better webmaster information than anything Google offered to the public -- and you don't just roll out something like this in a couple days.
[ysearchblog.com...]
did find this interesting
[developer.yahoo.com...]
It's a bit on the techie site but you can easily create your own searchs
[edited by: Bewenched at 3:46 am (utc) on Aug. 9, 2006]
Am I missing something - Crawl date and language are the only extra things we get right? (I know what language my own sites are in... not worth authgetnticating just for that!)
[mydomain.com...]
[mydomain.com...]
Only thing I can think of is that we're blocking Y! *somehow*. So, my question is... has anyone checked their logs to see how Y! is grabbing the authentication file? Is there a user-agent or IP address I can check to see if I've blocked them? We have some IP ranges blocked, but none jump out at me as being a culprit. We also have some spider UA's blocked via .htaccess, but nothing in our blocking logs shows any spiders/bots hitting the authentication files. I'm kinda baffled.
And Y! provides zero information as to why the authentication failed, of course :)
Is that it? I can't see anything like "last crawled" or what language my site is in. In any case both I know already anyway!
I know Yahoo is going to add more stuff to this tool but at the moment it's pretty useless. I will wait and see what they do with the list of deep links I submitted. I think I'll also make a few dummy pages not linked from anywhere else and submit those to see what Yahoo does with them.
Is that it? I can't see anything like "last crawled" or what language my site is in. In any case both I know already anyway!
Very powerful stuff.
When they hit my authentication file, it was from the IP address 72.30.76.134 using the user agent "curl/7.13.2 (i686-pc-linux-gnu) libcurl/7.13.2 OpenSSL/0.9.7g zlib/1.2.2"
Thanks - I think that was the culprit. We had libcurl sort of blocked, so we'll see if unblocking it does the trick (it didn't need to be blocking the auth file!).
It's installed on our site using the correct tailored code but all searches produce a "No results found page".
I've emailed them to see what the problem is.
I'd like this to work, but if it doesn't it will have to come off our site.
Any news Tim?
Bewenched: How many pages did you have prior to the submit? Are you >saying the number of pages went down? Interesting Tool.
Nope .. number went up and it's been staying the same for months now. So either it was a fluke or it went up just about the same number of links (maybe a few less) I put in the file.
We normally hover right at 40k, were now up +700 with a new crawl date of last night even on pages I didn't add to the list. I only added ones that I knew were really deep. Gotta parse that file and find out which ones are new against the log files.
I better go make some coffee.
[edited by: Bewenched at 12:31 am (utc) on Aug. 11, 2006]
They also seem to be checking the authkey file every few hours which seems a bit excessive.
Maybe in time the validating frequency will slow down. It does make sense to check the file frequently in the beginning in case someone gained illegitimate access to the server in question and the rightful owner deleted the file.