1970: Use milli reexported tokenizer  r=curquiza a=ManyTheFish

Use milli reexported tokenizer instead of importing meilisearch-tokenizer dependency.

fix #1888

Co-authored-by: many <maxime@meilisearch.com>
This commit is contained in:
bors[bot] 2021-12-07 08:50:44 +00:00 committed by GitHub
commit 82033f935e
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
4 changed files with 2 additions and 22 deletions

20
Cargo.lock generated
View File

@ -1685,7 +1685,6 @@ dependencies = [
"meilisearch-auth", "meilisearch-auth",
"meilisearch-error", "meilisearch-error",
"meilisearch-lib", "meilisearch-lib",
"meilisearch-tokenizer 0.2.5",
"mime", "mime",
"num_cpus", "num_cpus",
"obkv", "obkv",
@ -1751,7 +1750,6 @@ dependencies = [
"lazy_static", "lazy_static",
"log", "log",
"meilisearch-error", "meilisearch-error",
"meilisearch-tokenizer 0.2.5",
"milli", "milli",
"mime", "mime",
"mockall", "mockall",
@ -1784,22 +1782,6 @@ dependencies = [
"whoami", "whoami",
] ]
[[package]]
name = "meilisearch-tokenizer"
version = "0.2.5"
source = "git+https://github.com/meilisearch/tokenizer.git?tag=v0.2.5#c0b5cf741ed9485147f2cbe523f2214d4fa4c395"
dependencies = [
"character_converter",
"cow-utils",
"deunicode",
"fst",
"jieba-rs",
"once_cell",
"slice-group-by",
"unicode-segmentation",
"whatlang",
]
[[package]] [[package]]
name = "meilisearch-tokenizer" name = "meilisearch-tokenizer"
version = "0.2.6" version = "0.2.6"
@ -1867,7 +1849,7 @@ dependencies = [
"linked-hash-map", "linked-hash-map",
"log", "log",
"logging_timer", "logging_timer",
"meilisearch-tokenizer 0.2.6", "meilisearch-tokenizer",
"memmap2", "memmap2",
"obkv", "obkv",
"once_cell", "once_cell",

View File

@ -50,7 +50,6 @@ log = "0.4.14"
meilisearch-auth = { path = "../meilisearch-auth" } meilisearch-auth = { path = "../meilisearch-auth" }
meilisearch-error = { path = "../meilisearch-error" } meilisearch-error = { path = "../meilisearch-error" }
meilisearch-lib = { path = "../meilisearch-lib" } meilisearch-lib = { path = "../meilisearch-lib" }
meilisearch-tokenizer = { git = "https://github.com/meilisearch/tokenizer.git", tag = "v0.2.5" }
mime = "0.3.16" mime = "0.3.16"
num_cpus = "1.13.0" num_cpus = "1.13.0"
obkv = "0.2.0" obkv = "0.2.0"

View File

@ -29,7 +29,6 @@ itertools = "0.10.1"
lazy_static = "1.4.0" lazy_static = "1.4.0"
log = "0.4.14" log = "0.4.14"
meilisearch-error = { path = "../meilisearch-error" } meilisearch-error = { path = "../meilisearch-error" }
meilisearch-tokenizer = { git = "https://github.com/meilisearch/tokenizer.git", tag = "v0.2.5" }
milli = { git = "https://github.com/meilisearch/milli.git", tag = "v0.21.0" } milli = { git = "https://github.com/meilisearch/milli.git", tag = "v0.21.0" }
mime = "0.3.16" mime = "0.3.16"
num_cpus = "1.13.0" num_cpus = "1.13.0"

View File

@ -4,7 +4,7 @@ use std::time::Instant;
use either::Either; use either::Either;
use indexmap::IndexMap; use indexmap::IndexMap;
use meilisearch_tokenizer::{Analyzer, AnalyzerConfig, Token}; use milli::tokenizer::{Analyzer, AnalyzerConfig, Token};
use milli::{AscDesc, FieldId, FieldsIdsMap, Filter, MatchingWords, SortError}; use milli::{AscDesc, FieldId, FieldsIdsMap, Filter, MatchingWords, SortError};
use regex::Regex; use regex::Regex;
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};