Forum Moderators: open

Message Too Old, No Replies

Deepindex to roll out german and english index

Free early bird submissions

         

heini

10:45 am on Nov 20, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



French SE deepindex.com is in beta tests for rolling out german and english dbs.

The english site is already up and running:
[en.deepindex.com...]

The beta version will be online within the next few days. We invite you to submit your sites for free and test out the search abilities over the next six weeks. Our expected launch date for the new Engine is 27th December 2002.

Start for the free crawl on the UK side is to be expected in the second half of december and for germany at the end of january. Both servers will crawl free for 3 months and submissions will be opened in december for both servers.

volatilegx

2:40 pm on Jan 17, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I just saw DeepIndex crawling a site under the User-Agent "DeepIndex". It came from the IP 62.212.117.198 (deepindex.net1.nerim.net)

DeepIndex

2:52 pm on Jan 17, 2003 (gmt 0)

10+ Year Member



Hi,

your are giving "cloaking" adress on the forum? :-D

Yes this is our principal agent and IP adress. i would like to have some feedback on the behaviour of the bot. Is he polite?

We are actualy having our Deep "dance" and hope that DeepIndex Bot is polite with your site. He is normally always following robots.txt and robots meta, and if he should not do it correctly, or if he did gather things he shouldn't, than tell me the urls i should exclude... i'll do that quickly as possible.

Best regards to all :-)

Yidaki

12:31 pm on Jan 18, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



You might want to recheck the hints and examples i sticky mailed to you on Jan 5!?

<edit>Removed the first part of this my post ... sounded too much like a complain ... it isn't!<edit>

[edited by: Yidaki at 1:21 pm (utc) on Jan. 18, 2003]

DeepIndex

1:01 pm on Jan 18, 2003 (gmt 0)

10+ Year Member



Hi,

update is on course. Should be finished this we.

What'syour url?

regards

Yidaki

1:28 pm on Jan 18, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Deepindex, sorry if my post sounded like a flame. However, checking the examples would be a good idea.

I'll wait some more days and see how deepindex develops.

Good Luck!

BTW: my url is www.sorryicanttellyou.de, ... ;)

DeepIndex

1:38 pm on Jan 18, 2003 (gmt 0)

10+ Year Member



arf, i'm not bothered, neither angry ;-)... but i had a hard week ... and a much more harder weekend.

I checkout you're url but i didn't see your url. anyway.. i push it in now! The bot is on your site right now :-))

I love getting feedback (good one and the bad one) but i prefer of course good feedback :-D

Yidaki

5:23 pm on Feb 12, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



deepindex, after months of no luck submitting my site to deepindex and after watching the bad result quality increase, i decided to try again your se only if it's recommended at slashdot or at the WebmasterWorld start page. So it's up to you ... ;)

DeepIndex

7:02 pm on Feb 12, 2003 (gmt 0)

10+ Year Member



Hi

> deepindex, after months of no luck submitting

hmmm ... i'm not sure, but i think i included your site...as i told you on mail... do i have a bad memory?

Another thing to think about is to ask why a lot's of engines do disappear.

Is it because google is too dominant?

Is it because they are not given the chance to come up?

Is it because it's not perfect at the beginning and it's getting better with the time?

Is an SE project like DeepIndex bad because of some doggy insurance companies which did do some spamdexing... (even i do not believe that these are really spamdexed)

But the time costs money, a lot of money, you know that too, and for an SE it's even more money... than a directory... so give deepindex a chance, help us to get better ...

Yidaki

7:17 pm on Feb 12, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member




> hmmm ... i'm not sure, but i think i included your site

Nope... you said so but you didn't. And the results also didn't change and are still brute force spammed. Try the simple first example i sticky mailed you months ago and you'll get it.

I'm sorry but that aren't good signs. :(

DeepIndex

9:02 pm on Feb 12, 2003 (gmt 0)

10+ Year Member



Hmmm

most of the submissions receved where indexed by now. Two things are possible :

either i missed your site for a reason ... wrong devination of yoururl.com

either when i get to much submissions coming with the same domain i pushed your url out without wanting it...sorry

You gave me also some consideration on how to optimize to avoid spamdexing sites... most of them are surely options to consider, what i do, but as i experienced already on the french side of deepindex, the accuracy comes with the growing of the size of the index. When you look at great SE's like Google, you'll find also a lot's of pages who are pure spamdexing pages... depending mostly on competif search terms... so neither a very good google, neither a very bad deepindex can avoid this and we all do know that.

Now, the thing i don't like in the post is

>"I'm sorry but that aren't good signs. :("<

Installing a general purpose SE is much more difficult than to make specialized SE or Directory, where you are more selectif than an SE who does a full crawl, with following links on your sites and on others...

Yidaki

9:38 pm on Feb 12, 2003 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



DeepIndex, first thing: don't get me wrong. I don't complain about your project but as i said you before, there are some serious things you should do to improve your quality.

I understand that some algos work better with large indizes but some kinds of spam can easily be filtered even in small indizes. (Keyword stuffing and repetition, sneaky redirects, pron spam ...)

>Installing a general purpose SE is much more difficult than to make specialized SE or Directory

Nope! If you run a specialized se that doesn't crawl the web you might be right. But a specialized se that continuously crawls the web, weeds out the spam first and then takes the links that are related to the given topic can be even more complex than a general se.

IMHO filtering spam is a main part of any good se algo - search quality starts with this.

BTW: no, my submissions were not crawled by deepindex - there was no typo - and i only submitted two different domains. But i'll do it again ... ;)

This 41 message thread spans 2 pages: 41