| 1:15 pm on Jul 9, 2008 (gmt 0)|
It seems that your "leech turned on" is generating a redirect/replacement file (via 302).
This is really not an effective presentation for major SE's.
In fact, presenting alternative images to any visitor (i. e., leech) as opposed to a straight-forward access-forbidden (403) is a bad practice.
You need to modify the CP created htaccess to effective reso;ution for both standard visitors and major SE's.
| 12:32 am on Jul 10, 2008 (gmt 0)|
I don't think I have access to the .htaccess file. I'm on a shared server. Does that make a difference? It seems from discussions that some hosting services do allow this. I got an all-you-can-eat deal and I took them up on it :) for a low yearly price. Tho now I am tempted to switch as I've seen some good ideas in these discussions, if that's what it takes.
And why would it sometimes create a 404 and other times not?
I don't want people linking directly to my images as I have web pages, and it helps me ID scraper bots too if they try and grab jpegs with not being referred from my pages. Not too many are good enuf to follow links and get files that way.
| 12:47 pm on Jul 10, 2008 (gmt 0)|
|I don't think I have access to the .htaccess file. I'm on a shared server. Does that make a difference? It seems from discussions that some hosting services do allow this. |
You really need to poke around on your host/server site and see what's available.
Most likely if the CP (Control Panel) offers the "leech" than it also offers manual editing of your htaccess.
In my tenth year of utilizing Apache-shared hosting, I've never saw a provider that did not allow htaccess editing.
It could be different if your using and IIS server.
In addition, and in another thread, you noted that raw access logs were not available to you?
I would poke around until you locate them.
| 1:52 pm on Jul 10, 2008 (gmt 0)|
|In addition, and in another thread, you noted that raw access logs were not available to you? |
I would poke around until you locate them.
I'm in the same boat - no raw logs. Have had a pretty good poke around with both ftp and the cp file manager without any luck. Can you give any additional hints? The hosting company I'm using is also a big name in domains...
| 2:09 pm on Jul 10, 2008 (gmt 0)|
In March I switched the hosting on one of my domains to "El Cheapo" ;)
Although the initial transition and corrections required in my htaccess file were a nightmare, the host although with some occasional live problems is OK for me.
This host utilizes a service for support and although the support response is impressive, the laguage barriers create some real issues on each and every support request.
Each tech, requires a different terminology.
This host and others offer multiple domain hosting at the cheap price. I've never seen the "OTHERS" CP, however this one does not offer FTP access to raw logs (which is unheard of for me).
In addition the CP REQUIRES java on for full fucntions.
The Raw logs are located under CP:
1) Live Stats (pulldown)
2) Access & Error Logs
The host assured me prior to my 30-day money back guarantee expiriation that they offer "rotating logs", however unless they rotate on a monthly basis, I haven't seen daily rotating logs which I previously had grown accustomed to.
Were I to let my daily logs accumulate for a month, checking any access would be a nightmare (as would bandwidth use).
Thus, I manually delete and restart the logs every few days, losing a few hours accumulation in the process.
| 7:17 pm on Jul 10, 2008 (gmt 0)|
Thanks; I've tried to find the .htaccess file, nothing under that name in any of the root diretories (of shared server).
I DO have access to the raw logs.
I've had the ISP do really stupid stuff like lose the monthly archived raw access logs then tell me that I should use the canned control panel stats. These include robots which I want to exclude when looking for visitors. Analog is totally useless as it lists so many URL's (I have a humongo site) that it crashes the browser on most months. webalizer works but is limited. they just added awstats and it is a bit better in excluding robots but it is only based on UA with web page in it not behavior.
And they end up now resetting the raw logs DAILY!? I really should ask them about this. It used to be monthly.
However I have a script going in background that grabs the raw log hourly, separates out bots I know about (via a lookup table of either the full address, and if not found the 1st 3 groups of IP address).
Then the raw stats are written as is (skipping duplicate writes) to either visitors or bots database table.
From there is simply query to look up visitors, and if a new bot appears I just add it to the lookup table so it doesn't show up as a visitor again. (I put myself in bots table as a bot named 'me' or 'me2' if dift address at work so when we visit our site doesn't show up in visitor counts).
| 9:11 pm on Jul 10, 2008 (gmt 0)|
Something I didn't realize: being a bot, it is up to the bot to determine if it wants to actually request the 404 page, the same way it can skip requesting graphics and just grab html, and it appears sometimes it does and sometimes it doesn't want the 404 page apparently.
| 10:10 pm on Jul 10, 2008 (gmt 0)|
1) If the page doesn't exist, the bot has no choice.
2) with properly implemented custom error pages (server side), the bot has no choice either.
| 10:19 pm on Jul 10, 2008 (gmt 0)|
If you use PHP, you can get that to write a log of pages requested for you.
Google does use some of the 220.127.116.11 - 18.104.22.168 range for its own bots.
A bot will not have a referrer, because it does not jump from page to page following links. It spiders links from an internal list of what to spider, the link references collected some hours to days before on a previous run.
| 10:28 pm on Jul 10, 2008 (gmt 0)|
Could you possibbly expand on this for clarification that we are on the same channel?
1) Are you utilizing paid-webhosting which functions on under a shared server hosting plan [en.wikipedia.org]?
2) or perhaps some alternative/mis-configured plan offered to you via a friend or perhaps free host?
I've never seen a shared hosting server which did not offer use of htaccess, although I recall one of my hosts was without a file in place when my account my was new. Creating one was no real issue, however, I'm more proned to believe that you personally are simply unable to locate the file.
Rotaing logs are generally dumped into accumulating ZIP files.
The Windows version of Analog offers a config file for customized stats (the shared host stats are not specifically customized for you or your sites, rather the servers sites as a group), and as a result, there is simply no need to tear you daily logs apart from their original format, throwing a monkey wrench (so to speak) into the mix.
However, "to each his own".
My effort in providing these explanations to you is under the hope that your experiences at Webmaster World and those others that may assist you will require less communication in the future.
| 11:26 pm on Jul 10, 2008 (gmt 0)|
Most FTP applications are set up to "hide filenames that begin with a dot" by default.
Make sure you cancel that option in your FTP program, before you look round your site.
Err, that was meant for a thread talking about finding the .htaccess file on the server.
| 11:52 pm on Jul 10, 2008 (gmt 0)|
|A bot will not have a referrer |
Legit bots don't, but fakers trying to slip by simplistic anti-scraper rules include a referrer. Typically it's stupid and just uses your index page on all hits, even to pages that aren't even on the index page, but some really follow the flow of the site.
| 12:19 am on Jul 11, 2008 (gmt 0)|
|Most FTP applications are set up to "hide filenames that begin with a dot" by default. |
My "El Cheapo" host explained to me that under their CP file format, FTP access to log was not possible, without offering open http access to anybody to same logs.
Had I never seen ftp logs made available by a host above my own root folder, I may have accepted the explantion. Instead, I just shrugged and proclaimed to myself "CHEAP" ;)