mirror of
https://github.com/meilisearch/MeiliSearch
synced 2024-11-11 07:28:56 +01:00
1310 lines
64 KiB
Rust
1310 lines
64 KiB
Rust
use std::borrow::Cow;
|
|
use std::collections::HashMap;
|
|
use std::ops::{Range, Deref};
|
|
use std::time::Duration;
|
|
|
|
use either::Either;
|
|
use sdset::SetOperation;
|
|
|
|
use meilisearch_schema::FieldId;
|
|
|
|
use crate::database::MainT;
|
|
use crate::bucket_sort::{bucket_sort, bucket_sort_with_distinct, SortResult};
|
|
use crate::{criterion::Criteria, DocumentId};
|
|
use crate::{reordered_attrs::ReorderedAttrs, store, MResult};
|
|
use crate::facets::FacetFilter;
|
|
|
|
pub struct QueryBuilder<'c, 'f, 'd, 'i> {
|
|
criteria: Criteria<'c>,
|
|
searchable_attrs: Option<ReorderedAttrs>,
|
|
filter: Option<Box<dyn Fn(DocumentId) -> bool + 'f>>,
|
|
distinct: Option<(Box<dyn Fn(DocumentId) -> Option<u64> + 'd>, usize)>,
|
|
timeout: Option<Duration>,
|
|
index: &'i store::Index,
|
|
facet_filter: Option<FacetFilter>,
|
|
facets: Option<Vec<(FieldId, String)>>,
|
|
}
|
|
|
|
impl<'c, 'f, 'd, 'i> QueryBuilder<'c, 'f, 'd, 'i> {
|
|
pub fn new(index: &'i store::Index) -> Self {
|
|
QueryBuilder::with_criteria(
|
|
index,
|
|
Criteria::default(),
|
|
)
|
|
}
|
|
|
|
/// sets facet attributes to filter on
|
|
pub fn set_facet_filter(&mut self, facets: Option<FacetFilter>) {
|
|
self.facet_filter = facets;
|
|
}
|
|
|
|
/// sets facet attributes for which to return the count
|
|
pub fn set_facets(&mut self, facets: Option<Vec<(FieldId, String)>>) {
|
|
self.facets = facets;
|
|
}
|
|
|
|
pub fn with_criteria(
|
|
index: &'i store::Index,
|
|
criteria: Criteria<'c>,
|
|
) -> Self {
|
|
QueryBuilder {
|
|
criteria,
|
|
searchable_attrs: None,
|
|
filter: None,
|
|
distinct: None,
|
|
timeout: None,
|
|
index,
|
|
facet_filter: None,
|
|
facets: None,
|
|
}
|
|
}
|
|
|
|
pub fn with_filter<F>(&mut self, function: F)
|
|
where
|
|
F: Fn(DocumentId) -> bool + 'f,
|
|
{
|
|
self.filter = Some(Box::new(function))
|
|
}
|
|
|
|
pub fn with_fetch_timeout(&mut self, timeout: Duration) {
|
|
self.timeout = Some(timeout)
|
|
}
|
|
|
|
pub fn with_distinct<F>(&mut self, size: usize, function: F)
|
|
where
|
|
F: Fn(DocumentId) -> Option<u64> + 'd,
|
|
{
|
|
self.distinct = Some((Box::new(function), size))
|
|
}
|
|
|
|
pub fn add_searchable_attribute(&mut self, attribute: u16) {
|
|
let reorders = self.searchable_attrs.get_or_insert_with(ReorderedAttrs::new);
|
|
reorders.insert_attribute(attribute);
|
|
}
|
|
|
|
pub fn query(
|
|
self,
|
|
reader: &heed::RoTxn<MainT>,
|
|
query: &str,
|
|
range: Range<usize>,
|
|
) -> MResult<SortResult> {
|
|
let facets_docids = match self.facet_filter {
|
|
Some(facets) => {
|
|
let mut ands = Vec::with_capacity(facets.len());
|
|
let mut ors = Vec::new();
|
|
for f in facets.deref() {
|
|
match f {
|
|
Either::Left(keys) => {
|
|
ors.reserve(keys.len());
|
|
for key in keys {
|
|
let docids = self.index.facets.facet_document_ids(reader, &key)?.unwrap_or_default();
|
|
ors.push(docids);
|
|
}
|
|
let sets: Vec<_> = ors.iter().map(Cow::deref).collect();
|
|
let or_result = sdset::multi::OpBuilder::from_vec(sets).union().into_set_buf();
|
|
ands.push(Cow::Owned(or_result));
|
|
ors.clear();
|
|
}
|
|
Either::Right(key) =>{
|
|
match self.index.facets.facet_document_ids(reader, &key)? {
|
|
Some(docids) => ands.push(docids),
|
|
// no candidates for search, early return.
|
|
None => return Ok(SortResult::default()),
|
|
}
|
|
}
|
|
};
|
|
}
|
|
let ands: Vec<_> = ands.iter().map(Cow::deref).collect();
|
|
Some(sdset::multi::OpBuilder::from_vec(ands).intersection().into_set_buf())
|
|
}
|
|
None => None
|
|
};
|
|
|
|
// for each field to retrieve the count for, create an HashMap associating the attribute
|
|
// value to a set of matching documents. The HashMaps are them collected in another
|
|
// HashMap, associating each HashMap to it's field.
|
|
let facet_count_docids = match self.facets {
|
|
Some(field_ids) => {
|
|
let mut facet_count_map = HashMap::new();
|
|
for (field_id, field_name) in field_ids {
|
|
let mut key_map = HashMap::new();
|
|
for pair in self.index.facets.field_document_ids(reader, field_id)? {
|
|
let (facet_key, document_ids) = pair?;
|
|
let value = facet_key.value();
|
|
key_map.insert(value.to_string(), document_ids);
|
|
}
|
|
facet_count_map.insert(field_name, key_map);
|
|
}
|
|
Some(facet_count_map)
|
|
}
|
|
None => None,
|
|
};
|
|
|
|
match self.distinct {
|
|
Some((distinct, distinct_size)) => bucket_sort_with_distinct(
|
|
reader,
|
|
query,
|
|
range,
|
|
facets_docids,
|
|
facet_count_docids,
|
|
self.filter,
|
|
distinct,
|
|
distinct_size,
|
|
self.criteria,
|
|
self.searchable_attrs,
|
|
self.index.main,
|
|
self.index.postings_lists,
|
|
self.index.documents_fields_counts,
|
|
self.index.synonyms,
|
|
self.index.prefix_documents_cache,
|
|
self.index.prefix_postings_lists_cache,
|
|
),
|
|
None => bucket_sort(
|
|
reader,
|
|
query,
|
|
range,
|
|
facets_docids,
|
|
facet_count_docids,
|
|
self.filter,
|
|
self.criteria,
|
|
self.searchable_attrs,
|
|
self.index.main,
|
|
self.index.postings_lists,
|
|
self.index.documents_fields_counts,
|
|
self.index.synonyms,
|
|
self.index.prefix_documents_cache,
|
|
self.index.prefix_postings_lists_cache,
|
|
),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
|
|
use std::collections::{BTreeSet, HashMap};
|
|
use std::iter::FromIterator;
|
|
|
|
use fst::IntoStreamer;
|
|
use meilisearch_schema::IndexedPos;
|
|
use sdset::SetBuf;
|
|
use tempfile::TempDir;
|
|
|
|
use crate::DocIndex;
|
|
use crate::Document;
|
|
use crate::automaton::normalize_str;
|
|
use crate::bucket_sort::SimpleMatch;
|
|
use crate::database::{Database,DatabaseOptions};
|
|
use crate::store::Index;
|
|
use meilisearch_schema::Schema;
|
|
|
|
fn set_from_stream<'f, I, S>(stream: I) -> fst::Set<Vec<u8>>
|
|
where
|
|
I: for<'a> fst::IntoStreamer<'a, Into = S, Item = &'a [u8]>,
|
|
S: 'f + for<'a> fst::Streamer<'a, Item = &'a [u8]>,
|
|
{
|
|
let mut builder = fst::SetBuilder::memory();
|
|
builder.extend_stream(stream).unwrap();
|
|
builder.into_set()
|
|
}
|
|
|
|
fn insert_key<A: AsRef<[u8]>>(set: &fst::Set<A>, key: &[u8]) -> fst::Set<Vec<u8>> {
|
|
let unique_key = {
|
|
let mut builder = fst::SetBuilder::memory();
|
|
builder.insert(key).unwrap();
|
|
builder.into_set()
|
|
};
|
|
|
|
let union_ = set.op().add(unique_key.into_stream()).r#union();
|
|
|
|
set_from_stream(union_)
|
|
}
|
|
|
|
fn sdset_into_fstset(set: &sdset::Set<&str>) -> fst::Set<Vec<u8>> {
|
|
let mut builder = fst::SetBuilder::memory();
|
|
let set = SetBuf::from_dirty(set.into_iter().map(|s| normalize_str(s)).collect());
|
|
builder.extend_iter(set.into_iter()).unwrap();
|
|
builder.into_set()
|
|
}
|
|
|
|
const fn doc_index(document_id: u32, word_index: u16) -> DocIndex {
|
|
DocIndex {
|
|
document_id: DocumentId(document_id),
|
|
attribute: 0,
|
|
word_index,
|
|
char_index: 0,
|
|
char_length: 0,
|
|
}
|
|
}
|
|
|
|
const fn doc_char_index(document_id: u32, word_index: u16, char_index: u16) -> DocIndex {
|
|
DocIndex {
|
|
document_id: DocumentId(document_id),
|
|
attribute: 0,
|
|
word_index,
|
|
char_index,
|
|
char_length: 0,
|
|
}
|
|
}
|
|
|
|
pub struct TempDatabase {
|
|
database: Database,
|
|
index: Index,
|
|
_tempdir: TempDir,
|
|
}
|
|
|
|
impl TempDatabase {
|
|
pub fn query_builder(&self) -> QueryBuilder {
|
|
self.index.query_builder()
|
|
}
|
|
|
|
pub fn add_synonym(&mut self, word: &str, new: SetBuf<&str>) {
|
|
let db = &self.database;
|
|
let mut writer = db.main_write_txn().unwrap();
|
|
|
|
let word = normalize_str(word);
|
|
|
|
let alternatives = self
|
|
.index
|
|
.synonyms
|
|
.synonyms(&writer, word.as_bytes())
|
|
.unwrap();
|
|
|
|
let new = sdset_into_fstset(&new);
|
|
let new_alternatives =
|
|
set_from_stream(alternatives.op().add(new.into_stream()).r#union());
|
|
self.index
|
|
.synonyms
|
|
.put_synonyms(&mut writer, word.as_bytes(), &new_alternatives)
|
|
.unwrap();
|
|
|
|
let synonyms = self.index.main.synonyms_fst(&writer).unwrap();
|
|
|
|
let synonyms_fst = insert_key(&synonyms, word.as_bytes());
|
|
self.index
|
|
.main
|
|
.put_synonyms_fst(&mut writer, &synonyms_fst)
|
|
.unwrap();
|
|
|
|
writer.commit().unwrap();
|
|
}
|
|
}
|
|
|
|
impl<'a> FromIterator<(&'a str, &'a [DocIndex])> for TempDatabase {
|
|
fn from_iter<I: IntoIterator<Item = (&'a str, &'a [DocIndex])>>(iter: I) -> Self {
|
|
let tempdir = TempDir::new().unwrap();
|
|
let database = Database::open_or_create(&tempdir, DatabaseOptions::default()).unwrap();
|
|
let index = database.create_index("default").unwrap();
|
|
|
|
let db = &database;
|
|
let mut writer = db.main_write_txn().unwrap();
|
|
|
|
let mut words_fst = BTreeSet::new();
|
|
let mut postings_lists = HashMap::new();
|
|
let mut fields_counts = HashMap::<_, u16>::new();
|
|
|
|
let mut schema = Schema::with_primary_key("id");
|
|
|
|
for (word, indexes) in iter {
|
|
let mut final_indexes = Vec::new();
|
|
for index in indexes {
|
|
let name = index.attribute.to_string();
|
|
schema.insert(&name).unwrap();
|
|
let indexed_pos = schema.set_indexed(&name).unwrap().1;
|
|
let index = DocIndex {
|
|
attribute: indexed_pos.0,
|
|
..*index
|
|
};
|
|
final_indexes.push(index);
|
|
}
|
|
|
|
let word = word.to_lowercase().into_bytes();
|
|
words_fst.insert(word.clone());
|
|
postings_lists
|
|
.entry(word)
|
|
.or_insert_with(Vec::new)
|
|
.extend_from_slice(&final_indexes);
|
|
for idx in final_indexes {
|
|
fields_counts.insert((idx.document_id, idx.attribute, idx.word_index), 1);
|
|
}
|
|
}
|
|
|
|
index.main.put_schema(&mut writer, &schema).unwrap();
|
|
|
|
let words_fst = fst::Set::from_iter(words_fst).unwrap();
|
|
|
|
index.main.put_words_fst(&mut writer, &words_fst).unwrap();
|
|
|
|
for (word, postings_list) in postings_lists {
|
|
let postings_list = SetBuf::from_dirty(postings_list);
|
|
index
|
|
.postings_lists
|
|
.put_postings_list(&mut writer, &word, &postings_list)
|
|
.unwrap();
|
|
}
|
|
|
|
for ((docid, attr, _), count) in fields_counts {
|
|
let prev = index
|
|
.documents_fields_counts
|
|
.document_field_count(&writer, docid, IndexedPos(attr))
|
|
.unwrap();
|
|
|
|
let prev = prev.unwrap_or(0);
|
|
|
|
index
|
|
.documents_fields_counts
|
|
.put_document_field_count(&mut writer, docid, IndexedPos(attr), prev + count)
|
|
.unwrap();
|
|
}
|
|
|
|
writer.commit().unwrap();
|
|
|
|
TempDatabase { database, index, _tempdir: tempdir }
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn simple() {
|
|
let store = TempDatabase::from_iter(vec![
|
|
("iphone", &[doc_char_index(0, 0, 0)][..]),
|
|
("from", &[doc_char_index(0, 1, 1)][..]),
|
|
("apple", &[doc_char_index(0, 2, 2)][..]),
|
|
]);
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "iphone from apple", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, .. }));
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn simple_synonyms() {
|
|
let mut store = TempDatabase::from_iter(vec![("hello", &[doc_index(0, 0)][..])]);
|
|
|
|
store.add_synonym("bonjour", SetBuf::from_dirty(vec!["hello"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "hello", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "bonjour", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
// #[test]
|
|
// fn prefix_synonyms() {
|
|
// let mut store = TempDatabase::from_iter(vec![("hello", &[doc_index(0, 0)][..])]);
|
|
|
|
// store.add_synonym("bonjour", SetBuf::from_dirty(vec!["hello"]));
|
|
// store.add_synonym("salut", SetBuf::from_dirty(vec!["hello"]));
|
|
|
|
// let db = &store.database;
|
|
// let reader = db.main_read_txn().unwrap();
|
|
|
|
// let builder = store.query_builder();
|
|
// let results = builder.query(&reader, "sal", 0..20).unwrap();
|
|
// let mut iter = documents.into_iter();
|
|
|
|
// assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
// let mut matches = matches.into_iter();
|
|
// assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
// assert_matches!(matches.next(), None);
|
|
// });
|
|
// assert_matches!(iter.next(), None);
|
|
|
|
// let builder = store.query_builder();
|
|
// let results = builder.query(&reader, "bonj", 0..20).unwrap();
|
|
// let mut iter = documents.into_iter();
|
|
|
|
// assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
// let mut matches = matches.into_iter();
|
|
// assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
// assert_matches!(matches.next(), None);
|
|
// });
|
|
// assert_matches!(iter.next(), None);
|
|
|
|
// let builder = store.query_builder();
|
|
// let results = builder.query(&reader, "sal blabla", 0..20).unwrap();
|
|
// let mut iter = documents.into_iter();
|
|
|
|
// assert_matches!(iter.next(), None);
|
|
|
|
// let builder = store.query_builder();
|
|
// let results = builder.query(&reader, "bonj blabla", 0..20).unwrap();
|
|
// let mut iter = documents.into_iter();
|
|
|
|
// assert_matches!(iter.next(), None);
|
|
// }
|
|
|
|
// #[test]
|
|
// fn levenshtein_synonyms() {
|
|
// let mut store = TempDatabase::from_iter(vec![("hello", &[doc_index(0, 0)][..])]);
|
|
|
|
// store.add_synonym("salutation", SetBuf::from_dirty(vec!["hello"]));
|
|
|
|
// let db = &store.database;
|
|
// let reader = db.main_read_txn().unwrap();
|
|
|
|
// let builder = store.query_builder();
|
|
// let results = builder.query(&reader, "salutution", 0..20).unwrap();
|
|
// let mut iter = documents.into_iter();
|
|
|
|
// assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
// let mut matches = matches.into_iter();
|
|
// assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
// assert_matches!(matches.next(), None);
|
|
// });
|
|
// assert_matches!(iter.next(), None);
|
|
|
|
// let builder = store.query_builder();
|
|
// let results = builder.query(&reader, "saluttion", 0..20).unwrap();
|
|
// let mut iter = documents.into_iter();
|
|
|
|
// assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
// let mut matches = matches.into_iter();
|
|
// assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
// assert_matches!(matches.next(), None);
|
|
// });
|
|
// assert_matches!(iter.next(), None);
|
|
// }
|
|
|
|
#[test]
|
|
fn harder_synonyms() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("hello", &[doc_index(0, 0)][..]),
|
|
("bonjour", &[doc_index(1, 3)]),
|
|
("salut", &[doc_index(2, 5)]),
|
|
]);
|
|
|
|
store.add_synonym("hello", SetBuf::from_dirty(vec!["bonjour", "salut"]));
|
|
store.add_synonym("bonjour", SetBuf::from_dirty(vec!["hello", "salut"]));
|
|
store.add_synonym("salut", SetBuf::from_dirty(vec!["hello", "bonjour"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "hello", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 3, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 5, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "bonjour", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 3, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 5, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "salut", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 3, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 5, .. }));
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
/// Unique word has multi-word synonyms
|
|
fn unique_to_multiword_synonyms() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("new", &[doc_char_index(0, 0, 0)][..]),
|
|
("york", &[doc_char_index(0, 1, 1)][..]),
|
|
("city", &[doc_char_index(0, 2, 2)][..]),
|
|
("subway", &[doc_char_index(0, 3, 3)][..]),
|
|
("NY", &[doc_char_index(1, 0, 0)][..]),
|
|
("subway", &[doc_char_index(1, 1, 1)][..]),
|
|
]);
|
|
|
|
store.add_synonym(
|
|
"NY",
|
|
SetBuf::from_dirty(vec!["NYC", "new york", "new york city"]),
|
|
);
|
|
store.add_synonym(
|
|
"NYC",
|
|
SetBuf::from_dirty(vec!["NY", "new york", "new york city"]),
|
|
);
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "NY subway", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // NY ± new
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // NY ± york
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // NY ± city
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult { documents, .. } = builder.query(&reader, "NYC subway", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // NYC ± new
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // NYC ± york
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // NYC ± city
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn unique_to_multiword_synonyms_words_proximity() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("new", &[doc_char_index(0, 0, 0)][..]),
|
|
("york", &[doc_char_index(0, 1, 1)][..]),
|
|
("city", &[doc_char_index(0, 2, 2)][..]),
|
|
("subway", &[doc_char_index(0, 3, 3)][..]),
|
|
("york", &[doc_char_index(1, 0, 0)][..]),
|
|
("new", &[doc_char_index(1, 1, 1)][..]),
|
|
("subway", &[doc_char_index(1, 2, 2)][..]),
|
|
("NY", &[doc_char_index(2, 0, 0)][..]),
|
|
("subway", &[doc_char_index(2, 1, 1)][..]),
|
|
]);
|
|
|
|
store.add_synonym("NY", SetBuf::from_dirty(vec!["york new"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "NY", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. })); // NY ± york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, .. })); // NY ± new
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. })); // york = NY
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, .. })); // new = NY
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 1, .. })); // york = NY
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 0, .. })); // new = NY
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "new york", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, .. })); // york
|
|
assert_matches!(matches.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 1, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 0, .. })); // new
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn unique_to_multiword_synonyms_cumulative_word_index() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("NY", &[doc_char_index(0, 0, 0)][..]),
|
|
("subway", &[doc_char_index(0, 1, 1)][..]),
|
|
("new", &[doc_char_index(1, 0, 0)][..]),
|
|
("york", &[doc_char_index(1, 1, 1)][..]),
|
|
("subway", &[doc_char_index(1, 2, 2)][..]),
|
|
]);
|
|
|
|
store.add_synonym("new york", SetBuf::from_dirty(vec!["NY"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "NY subway", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // NY
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // subway
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
// assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
// let mut matches = matches.into_iter();
|
|
// assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 2, is_exact: true, .. })); // subway
|
|
// assert_matches!(matches.next(), None);
|
|
// });
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "new york subway", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // subway
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new = NY
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york = NY
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // subway
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
/// Unique word has multi-word synonyms
|
|
fn harder_unique_to_multiword_synonyms_one() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("new", &[doc_char_index(0, 0, 0)][..]),
|
|
("york", &[doc_char_index(0, 1, 1)][..]),
|
|
("city", &[doc_char_index(0, 2, 2)][..]),
|
|
("yellow", &[doc_char_index(0, 3, 3)][..]),
|
|
("subway", &[doc_char_index(0, 4, 4)][..]),
|
|
("broken", &[doc_char_index(0, 5, 5)][..]),
|
|
("NY", &[doc_char_index(1, 0, 0)][..]),
|
|
("blue", &[doc_char_index(1, 1, 1)][..]),
|
|
("subway", &[doc_char_index(1, 2, 2)][..]),
|
|
]);
|
|
|
|
store.add_synonym(
|
|
"NY",
|
|
SetBuf::from_dirty(vec!["NYC", "new york", "new york city"]),
|
|
);
|
|
store.add_synonym(
|
|
"NYC",
|
|
SetBuf::from_dirty(vec!["NY", "new york", "new york city"]),
|
|
);
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "NY subway", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // new = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // city = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "NYC subway", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // NYC
|
|
// because one-word to one-word ^^^^
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // new = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // city = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: true, .. })); // subway
|
|
assert_matches!(iter.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
/// Unique word has multi-word synonyms
|
|
fn even_harder_unique_to_multiword_synonyms() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("new", &[doc_char_index(0, 0, 0)][..]),
|
|
("york", &[doc_char_index(0, 1, 1)][..]),
|
|
("city", &[doc_char_index(0, 2, 2)][..]),
|
|
("yellow", &[doc_char_index(0, 3, 3)][..]),
|
|
("underground", &[doc_char_index(0, 4, 4)][..]),
|
|
("train", &[doc_char_index(0, 5, 5)][..]),
|
|
("broken", &[doc_char_index(0, 6, 6)][..]),
|
|
("NY", &[doc_char_index(1, 0, 0)][..]),
|
|
("blue", &[doc_char_index(1, 1, 1)][..]),
|
|
("subway", &[doc_char_index(1, 2, 2)][..]),
|
|
]);
|
|
|
|
store.add_synonym(
|
|
"NY",
|
|
SetBuf::from_dirty(vec!["NYC", "new york", "new york city"]),
|
|
);
|
|
store.add_synonym(
|
|
"NYC",
|
|
SetBuf::from_dirty(vec!["NY", "new york", "new york city"]),
|
|
);
|
|
store.add_synonym("subway", SetBuf::from_dirty(vec!["underground train"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "NY subway broken", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // new = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // city = NY
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: false, .. })); // underground = subway
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 4, word_index: 5, is_exact: false, .. })); // train = subway
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 5, word_index: 6, is_exact: true, .. })); // broken
|
|
assert_matches!(iter.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "NYC subway", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: true, .. })); // underground = subway
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 4, word_index: 5, is_exact: true, .. })); // train = subway
|
|
assert_matches!(iter.next(), None); // position rewritten ^
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // new = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york = NYC
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // city = NYC
|
|
// because one-word to one-word ^^^^
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: false, .. })); // subway = underground
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 4, word_index: 5, is_exact: false, .. })); // subway = train
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
/// Multi-word has multi-word synonyms
|
|
fn multiword_to_multiword_synonyms() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("NY", &[doc_char_index(0, 0, 0)][..]),
|
|
("subway", &[doc_char_index(0, 1, 1)][..]),
|
|
("NYC", &[doc_char_index(1, 0, 0)][..]),
|
|
("blue", &[doc_char_index(1, 1, 1)][..]),
|
|
("subway", &[doc_char_index(1, 2, 2)][..]),
|
|
("broken", &[doc_char_index(1, 3, 3)][..]),
|
|
("new", &[doc_char_index(2, 0, 0)][..]),
|
|
("york", &[doc_char_index(2, 1, 1)][..]),
|
|
("underground", &[doc_char_index(2, 2, 2)][..]),
|
|
("train", &[doc_char_index(2, 3, 3)][..]),
|
|
("broken", &[doc_char_index(2, 4, 4)][..]),
|
|
]);
|
|
|
|
store.add_synonym(
|
|
"new york",
|
|
SetBuf::from_dirty(vec!["NYC", "NY", "new york city"]),
|
|
);
|
|
store.add_synonym(
|
|
"new york city",
|
|
SetBuf::from_dirty(vec!["NYC", "NY", "new york"]),
|
|
);
|
|
store.add_synonym("underground train", SetBuf::from_dirty(vec!["subway"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder
|
|
.query(&reader, "new york underground train broken", 0..20)
|
|
.unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // underground
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 4, word_index: 4, is_exact: true, .. })); // train
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 5, word_index: 5, is_exact: true, .. })); // broken
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // NYC = new
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // NYC = york
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // NYC = city
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 4, is_exact: true, .. })); // subway = underground
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 4, word_index: 5, is_exact: true, .. })); // subway = train
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 5, word_index: 6, is_exact: true, .. })); // broken
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder
|
|
.query(&reader, "new york city underground train broken", 0..20)
|
|
.unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 3, word_index: 2, is_exact: true, .. })); // underground
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 4, word_index: 3, is_exact: true, .. })); // train
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 5, word_index: 4, is_exact: true, .. })); // broken
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // NYC = new
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // NYC = york
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // subway = underground
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 4, word_index: 4, is_exact: true, .. })); // subway = train
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 5, word_index: 5, is_exact: true, .. })); // broken
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn intercrossed_multiword_synonyms() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("new", &[doc_index(0, 0)][..]),
|
|
("york", &[doc_index(0, 1)][..]),
|
|
("big", &[doc_index(0, 2)][..]),
|
|
("city", &[doc_index(0, 3)][..]),
|
|
]);
|
|
|
|
store.add_synonym("new york", SetBuf::from_dirty(vec!["new york city"]));
|
|
store.add_synonym("new york city", SetBuf::from_dirty(vec!["new york"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "new york big ", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: false, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 4, is_exact: false, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // big
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("NY", &[doc_index(0, 0)][..]),
|
|
("city", &[doc_index(0, 1)][..]),
|
|
("subway", &[doc_index(0, 2)][..]),
|
|
("NY", &[doc_index(1, 0)][..]),
|
|
("subway", &[doc_index(1, 1)][..]),
|
|
("NY", &[doc_index(2, 0)][..]),
|
|
("york", &[doc_index(2, 1)][..]),
|
|
("city", &[doc_index(2, 2)][..]),
|
|
("subway", &[doc_index(2, 3)][..]),
|
|
]);
|
|
|
|
store.add_synonym("NY", SetBuf::from_dirty(vec!["new york city story"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "NY subway ", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 4, word_index: 3, is_exact: true, .. })); // subway
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: false, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: false, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 4, word_index: 3, is_exact: true, .. })); // subway
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // story
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 4, word_index: 4, is_exact: true, .. })); // subway
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn cumulative_word_indices() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("NYC", &[doc_index(0, 0)][..]),
|
|
("long", &[doc_index(0, 1)][..]),
|
|
("subway", &[doc_index(0, 2)][..]),
|
|
("cool", &[doc_index(0, 3)][..]),
|
|
]);
|
|
|
|
store.add_synonym("new york city", SetBuf::from_dirty(vec!["NYC"]));
|
|
store.add_synonym("subway", SetBuf::from_dirty(vec!["underground train"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder
|
|
.query(&reader, "new york city long subway cool ", 0..20)
|
|
.unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut matches = matches.into_iter();
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 0, word_index: 0, is_exact: true, .. })); // new = NYC
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 1, word_index: 1, is_exact: true, .. })); // york = NYC
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 2, word_index: 2, is_exact: true, .. })); // city = NYC
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 3, word_index: 3, is_exact: true, .. })); // long
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 4, word_index: 4, is_exact: true, .. })); // subway = underground
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 5, word_index: 5, is_exact: true, .. })); // subway = train
|
|
assert_matches!(matches.next(), Some(SimpleMatch { query_index: 6, word_index: 6, is_exact: true, .. })); // cool
|
|
assert_matches!(matches.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn deunicoded_synonyms() {
|
|
let mut store = TempDatabase::from_iter(vec![
|
|
("telephone", &[doc_index(0, 0)][..]), // meilisearch indexes the unidecoded
|
|
("téléphone", &[doc_index(0, 0)][..]), // and the original words on the same DocIndex
|
|
("iphone", &[doc_index(1, 0)][..]),
|
|
]);
|
|
|
|
store.add_synonym("téléphone", SetBuf::from_dirty(vec!["iphone"]));
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "telephone", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, .. }));
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, .. }));
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, .. }));
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "téléphone", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, .. }));
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, .. }));
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, .. }));
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "télephone", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, .. }));
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, distance: 1, word_index: 0, is_exact: false, .. })); // iphone | telephone
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn simple_concatenation() {
|
|
let store = TempDatabase::from_iter(vec![
|
|
("iphone", &[doc_index(0, 0)][..]),
|
|
("case", &[doc_index(0, 1)][..]),
|
|
]);
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "i phone case", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, distance: 0, .. })); // iphone
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 1, distance: 0, .. })); // iphone
|
|
// assert_matches!(iter.next(), Some(SimpleMatch { query_index: 1, word_index: 0, distance: 1, .. })); "phone"
|
|
// but no typo on first letter ^^^^^^^
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 2, word_index: 2, distance: 0, .. })); // case
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn exact_field_count_one_word() {
|
|
let store = TempDatabase::from_iter(vec![
|
|
("searchengine", &[doc_index(0, 0)][..]),
|
|
("searchengine", &[doc_index(1, 0)][..]),
|
|
("blue", &[doc_index(1, 1)][..]),
|
|
("searchangine", &[doc_index(2, 0)][..]),
|
|
("searchengine", &[doc_index(3, 0)][..]),
|
|
]);
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "searchengine", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, distance: 0, .. })); // searchengine
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(3), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, distance: 0, .. })); // searchengine
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, distance: 0, .. })); // searchengine
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(2), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, distance: 1, .. })); // searchengine
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn simple_phrase_query_splitting() {
|
|
let store = TempDatabase::from_iter(vec![
|
|
("search", &[doc_index(0, 0)][..]),
|
|
("engine", &[doc_index(0, 1)][..]),
|
|
("search", &[doc_index(1, 0)][..]),
|
|
("slow", &[doc_index(1, 1)][..]),
|
|
("engine", &[doc_index(1, 2)][..]),
|
|
]);
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "searchengine", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 0, distance: 0, .. })); // search
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 1, distance: 0, .. })); // engine
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
|
|
#[test]
|
|
fn harder_phrase_query_splitting() {
|
|
let store = TempDatabase::from_iter(vec![
|
|
("search", &[doc_index(0, 0)][..]),
|
|
("search", &[doc_index(0, 1)][..]),
|
|
("engine", &[doc_index(0, 2)][..]),
|
|
("search", &[doc_index(1, 0)][..]),
|
|
("slow", &[doc_index(1, 1)][..]),
|
|
("search", &[doc_index(1, 2)][..]),
|
|
("engine", &[doc_index(1, 3)][..]),
|
|
("search", &[doc_index(1, 0)][..]),
|
|
("search", &[doc_index(1, 1)][..]),
|
|
("slow", &[doc_index(1, 2)][..]),
|
|
("engine", &[doc_index(1, 3)][..]),
|
|
]);
|
|
|
|
let db = &store.database;
|
|
let reader = db.main_read_txn().unwrap();
|
|
|
|
let builder = store.query_builder();
|
|
let SortResult {documents, .. } = builder.query(&reader, "searchengine", 0..20).unwrap();
|
|
let mut iter = documents.into_iter();
|
|
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(0), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 1, distance: 0, .. })); // search
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 2, distance: 0, .. })); // engine
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), Some(Document { id: DocumentId(1), matches, .. }) => {
|
|
let mut iter = matches.into_iter();
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 2, distance: 0, .. })); // search
|
|
assert_matches!(iter.next(), Some(SimpleMatch { query_index: 0, word_index: 3, distance: 0, .. })); // engine
|
|
assert_matches!(iter.next(), None);
|
|
});
|
|
assert_matches!(iter.next(), None);
|
|
}
|
|
}
|