aa1ce97748
317: Fix the facet string docids filterable deletion bug r=Kerollmops a=Kerollmops Fixes a bug where the deletion of documents was returning a decoding error. But only when the settings are set with filterable attributes. This bug was introduced in #254 in which we made the engine faster in returning the facet distribution. We changed the way we were storing the inverted index, we were no more storing only documents ids with the original values but also groups identified with integers, depending on the facet level we were using. This is similar to how facet numbers are already stored. ⚠️ As `@curquiza` already said, we must first revert #309 before merging this! Related to https://github.com/meilisearch/MeiliSearch/issues/1601. Co-authored-by: Clément Renault <clement@meilisearch.com> |
||
---|---|---|
.github | ||
benchmarks | ||
helpers | ||
http-ui | ||
infos | ||
milli | ||
script | ||
search | ||
.gitignore | ||
.rustfmt.toml | ||
bors.toml | ||
Cargo.toml | ||
LICENSE | ||
README.md |
a concurrent indexer combined with fast and relevant search algorithms
Introduction
This repository contains the core engine used in MeiliSearch.
It contains a library that can manage one and only one index. MeiliSearch manages the multi-index itself. Milli is unable to store updates in a store: it is the job of something else above and this is why it is only able to process one update at a time.
This repository contains crates to quickly debug the engine:
- There are benchmarks located in the
benchmarks
crate. - The
http-ui
crate is a simple HTTP dashboard to tests the features like for real! - The
infos
crate is used to dump the internal data-structure and ensure correctness. - The
search
crate is a simple command-line that helps run flamegraph on top of it. - The
helpers
crate is only used to modify the database inplace, sometimes.
Compile and run the HTTP debug server
You can specify the number of threads to use to index documents and many other settings too.
cd http-ui
cargo run --release -- --db my-database.mdb -vvv --indexing-jobs 8
Index your documents
It can index a massive amount of documents in not much time, I already achieved to index:
- 115m songs (song and artist name) in ~1h and take 107GB on disk.
- 12m cities (name, timezone and country ID) in 15min and take 10GB on disk.
All of that on a 39$/month machine with 4cores.
You can feed the engine with your CSV (comma-seperated, yes) data like this:
printf "name,age\nhello,32\nkiki,24\n" | http POST 127.0.0.1:9700/documents content-type:text/csv
Here ids will be automatically generated as UUID v4 if they doesn't exist in some or every documents.
Note that it also support JSON and JSON streaming, you can send them to the engine by using
the content-type:application/json
and content-type:application/x-ndjson
headers respectively.
Querying the engine via the website
You can query the engine by going to the HTML page itself.
Contributing
You can setup a git-hook
to stop you from making a commit too fast. It'll stop you if:
- Any of the workspaces does not build
- Your code is not well-formatted
These two things are also checked in the CI, so ignoring the hook won't help you merge your code.
But if you need to, you can still add --no-verify
when creating your commit to ignore the hook.
To enable the hook, run the following command from the root of the project:
cp script/pre-commit .git/hooks/pre-commit