mirror of
https://github.com/meilisearch/MeiliSearch
synced 2024-11-04 20:18:55 +01:00
13f1277637
In PR #2773, I added the `chinese`, `hebrew`, `japanese` and `thai` feature flags to allow melisearch to be built without huge specialed tokenizations that took up 90% of the melisearch binary size. Unfortunately, due to some recent changes, this doesn't work anymore. The problem lies in excessive use of the `default` feature flag, which infects the dependency graph. Instead of adding `default-features = false` here and there, it's easier and more future-proof to not declare `default` in `milli` and `meilisearch-types`. I've renamed it to `all-tokenizers`, which also makes it a bit clearer what it's about.
56 lines
1.4 KiB
TOML
56 lines
1.4 KiB
TOML
[package]
|
|
name = "meilisearch-types"
|
|
publish = false
|
|
|
|
version.workspace = true
|
|
authors.workspace = true
|
|
description.workspace = true
|
|
homepage.workspace = true
|
|
readme.workspace = true
|
|
edition.workspace = true
|
|
license.workspace = true
|
|
|
|
[dependencies]
|
|
actix-web = { version = "4.3.1", default-features = false }
|
|
anyhow = "1.0.70"
|
|
convert_case = "0.6.0"
|
|
csv = "1.2.1"
|
|
deserr = "0.5.0"
|
|
either = { version = "1.8.1", features = ["serde"] }
|
|
enum-iterator = "1.4.0"
|
|
file-store = { path = "../file-store" }
|
|
flate2 = "1.0.25"
|
|
fst = "0.4.7"
|
|
memmap2 = "0.5.10"
|
|
milli = { path = "../milli" }
|
|
roaring = { version = "0.10.1", features = ["serde"] }
|
|
serde = { version = "1.0.160", features = ["derive"] }
|
|
serde-cs = "0.2.4"
|
|
serde_json = "1.0.95"
|
|
tar = "0.4.38"
|
|
tempfile = "3.5.0"
|
|
thiserror = "1.0.40"
|
|
time = { version = "0.3.20", features = ["serde-well-known", "formatting", "parsing", "macros"] }
|
|
tokio = "1.27"
|
|
uuid = { version = "1.3.1", features = ["serde", "v4"] }
|
|
|
|
[dev-dependencies]
|
|
insta = "1.29.0"
|
|
meili-snap = { path = "../meili-snap" }
|
|
|
|
[features]
|
|
# all specialized tokenizations
|
|
all-tokenizations = ["milli/all-tokenizations"]
|
|
|
|
# chinese specialized tokenization
|
|
chinese = ["milli/chinese"]
|
|
# hebrew specialized tokenization
|
|
hebrew = ["milli/hebrew"]
|
|
# japanese specialized tokenization
|
|
japanese = ["milli/japanese"]
|
|
# thai specialized tokenization
|
|
thai = ["milli/thai"]
|
|
|
|
# allow greek specialized tokenization
|
|
greek = ["milli/greek"]
|