Go to file
mpostma 3b60432687
Use update_id in UpdateBuilder
Add `the update_id` to the to the updates. The rationale is the
following:
- It allows for better tracability of the update events, thus improved
  debugging and logging.
- The enigne is now aware of what he's already processed, and can return
  it if asked. It may not make sense now, but in the future, the update
  store may not work the same way, and this information about the state
  of the engine will be desirable (distributed environement).
2021-02-01 19:46:34 +01:00
.github/workflows Avoid compiling benchmarks and speedup the CI 2020-11-11 11:14:57 +01:00
benches Fix the benchmarks 2020-10-31 22:18:29 +01:00
http-ui Use update_id in UpdateBuilder 2021-02-01 19:46:34 +01:00
src Use update_id in UpdateBuilder 2021-02-01 19:46:34 +01:00
.gitignore Move the http server into its own sub-module 2020-11-05 11:16:39 +01:00
Cargo.lock Remove the ordered-float serde feature 2021-01-27 14:11:10 +01:00
Cargo.toml Remove the ordered-float serde feature 2021-01-27 14:11:10 +01:00
LICENSE Initial commit 2020-05-31 14:21:56 +02:00
qc_loop.sh Initial commit 2020-05-31 14:22:06 +02:00
README.md Fix the milli logo in the README 2020-11-05 11:43:47 +01:00

the milli logo

a concurrent indexer combined with fast and relevant search algorithms

Introduction

This engine is a prototype, do not use it in production. This is one of the most advanced search engine I have worked on. It currently only supports the proximity criterion.

Compile and Run the server

You can specify the number of threads to use to index documents and many other settings too.

cd http-ui
cargo run --release -- serve --db my-database.mdb -vvv --indexing-jobs 8

Index your documents

It can index a massive amount of documents in not much time, I already achieved to index:

  • 115m songs (song and artist name) in ~1h and take 107GB on disk.
  • 12m cities (name, timezone and country ID) in 15min and take 10GB on disk.

All of that on a 39$/month machine with 4cores.

You can feed the engine with your CSV (comma-seperated, yes) data like this:

cat "name,age\nhello,32\nkiki,24\n" | http POST 127.0.0.1:9700/documents content-type:text/csv

Here ids will be automatically generated as UUID v4 if they doesn't exist in some or every documents.

Note that it also support JSON and JSON streaming, you can send them to the engine by using the content-type:application/json and content-type:application/x-ndjson headers respectively.

Querying the engine via the website

You can query the engine by going to the HTML page itself.