2023-03-15 16:08:43 +01:00
|
|
|
use std::collections::HashSet;
|
2023-02-21 09:42:54 +01:00
|
|
|
use std::mem;
|
|
|
|
use std::ops::RangeInclusive;
|
|
|
|
|
|
|
|
use charabia::normalizer::NormalizedTokenIter;
|
|
|
|
use charabia::{SeparatorKind, TokenKind};
|
|
|
|
use fst::automaton::Str;
|
|
|
|
use fst::{Automaton, IntoStreamer, Streamer};
|
|
|
|
use heed::types::DecodeIgnore;
|
|
|
|
use heed::RoTxn;
|
2023-03-02 21:27:57 +01:00
|
|
|
use itertools::Itertools;
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-14 16:37:47 +01:00
|
|
|
use super::interner::{DedupInterner, Interned};
|
2023-03-08 09:55:53 +01:00
|
|
|
use super::SearchContext;
|
2023-02-21 09:42:54 +01:00
|
|
|
use crate::search::fst_utils::{Complement, Intersection, StartsWith, Union};
|
|
|
|
use crate::search::{build_dfa, get_first};
|
2023-03-13 17:21:29 +01:00
|
|
|
use crate::{CboRoaringBitmapLenCodec, Index, Result, MAX_WORD_LENGTH};
|
2023-03-02 21:27:57 +01:00
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
/// A phrase in the user's search query, consisting of several words
|
|
|
|
/// that must appear side-by-side in the search results.
|
2023-03-06 19:21:55 +01:00
|
|
|
#[derive(Default, Clone, PartialEq, Eq, Hash)]
|
2023-03-02 21:27:57 +01:00
|
|
|
pub struct Phrase {
|
2023-03-06 19:21:55 +01:00
|
|
|
pub words: Vec<Option<Interned<String>>>,
|
2023-03-02 21:27:57 +01:00
|
|
|
}
|
|
|
|
impl Phrase {
|
2023-03-14 16:37:47 +01:00
|
|
|
pub fn description(&self, interner: &DedupInterner<String>) -> String {
|
2023-03-06 19:21:55 +01:00
|
|
|
self.words.iter().flatten().map(|w| interner.get(*w)).join(" ")
|
2023-03-02 21:27:57 +01:00
|
|
|
}
|
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
/// A structure storing all the different ways to match
|
|
|
|
/// a term in the user's search query.
|
2023-03-09 11:12:31 +01:00
|
|
|
#[derive(Clone, PartialEq, Eq, Hash)]
|
2023-03-14 10:54:55 +01:00
|
|
|
pub struct QueryTerm {
|
2023-03-13 17:21:29 +01:00
|
|
|
/// The original terms, for debugging purposes
|
2023-03-06 19:21:55 +01:00
|
|
|
pub original: Interned<String>,
|
2023-03-14 10:54:55 +01:00
|
|
|
/// Whether the term is an ngram
|
|
|
|
pub is_ngram: bool,
|
|
|
|
/// Whether the term can be only the prefix of a word
|
2023-03-13 17:21:29 +01:00
|
|
|
pub is_prefix: bool,
|
2023-03-14 10:54:55 +01:00
|
|
|
/// The original phrase, if any
|
|
|
|
pub phrase: Option<Interned<Phrase>>,
|
|
|
|
/// A single word equivalent to the original term, with zero typos
|
2023-03-13 17:21:29 +01:00
|
|
|
pub zero_typo: Option<Interned<String>>,
|
|
|
|
/// All the words that contain the original word as prefix
|
|
|
|
pub prefix_of: Box<[Interned<String>]>,
|
2023-03-14 10:54:55 +01:00
|
|
|
/// All the synonyms of the original word or phrase
|
2023-03-06 19:21:55 +01:00
|
|
|
pub synonyms: Box<[Interned<Phrase>]>,
|
2023-03-08 15:04:25 +01:00
|
|
|
|
|
|
|
/// The original word split into multiple consecutive words
|
2023-03-06 19:21:55 +01:00
|
|
|
pub split_words: Option<Interned<Phrase>>,
|
2023-03-08 15:04:25 +01:00
|
|
|
|
|
|
|
/// Words that are 1 typo away from the original word
|
2023-03-06 19:21:55 +01:00
|
|
|
pub one_typo: Box<[Interned<String>]>,
|
2023-03-08 15:04:25 +01:00
|
|
|
|
|
|
|
/// Words that are 2 typos away from the original word
|
2023-03-06 19:21:55 +01:00
|
|
|
pub two_typos: Box<[Interned<String>]>,
|
2023-03-08 15:04:25 +01:00
|
|
|
|
2023-03-13 17:21:29 +01:00
|
|
|
/// A prefix in the prefix databases matching the original word
|
|
|
|
pub use_prefix_db: Option<Interned<String>>,
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-14 10:54:55 +01:00
|
|
|
impl QueryTerm {
|
2023-03-15 16:08:43 +01:00
|
|
|
pub fn removing_forbidden_terms(
|
|
|
|
&self,
|
|
|
|
allowed_words: &HashSet<Interned<String>>,
|
|
|
|
allowed_phrases: &HashSet<Interned<Phrase>>,
|
|
|
|
) -> Option<Self> {
|
|
|
|
let QueryTerm {
|
|
|
|
original,
|
|
|
|
is_ngram,
|
|
|
|
is_prefix,
|
|
|
|
phrase,
|
|
|
|
zero_typo,
|
|
|
|
prefix_of,
|
|
|
|
synonyms,
|
|
|
|
split_words,
|
|
|
|
one_typo,
|
|
|
|
two_typos,
|
|
|
|
use_prefix_db,
|
|
|
|
} = self;
|
|
|
|
|
|
|
|
let mut changed = false;
|
|
|
|
|
|
|
|
let mut new_zero_typo = None;
|
|
|
|
if let Some(w) = zero_typo {
|
|
|
|
if allowed_words.contains(w) {
|
|
|
|
new_zero_typo = Some(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// TODO: this is incorrect, prefix DB stuff should be treated separately
|
|
|
|
let mut new_use_prefix_db = None;
|
|
|
|
if let Some(w) = use_prefix_db {
|
|
|
|
if allowed_words.contains(w) {
|
|
|
|
new_use_prefix_db = Some(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
let mut new_prefix_of = vec![];
|
|
|
|
for w in prefix_of.iter() {
|
|
|
|
if allowed_words.contains(w) {
|
|
|
|
new_prefix_of.push(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
let mut new_one_typo = vec![];
|
|
|
|
for w in one_typo.iter() {
|
|
|
|
if allowed_words.contains(w) {
|
|
|
|
new_one_typo.push(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
let mut new_two_typos = vec![];
|
|
|
|
for w in two_typos.iter() {
|
|
|
|
if allowed_words.contains(w) {
|
|
|
|
new_two_typos.push(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// TODO: this is incorrect, prefix DB stuff should be treated separately
|
|
|
|
let mut new_phrase = None;
|
|
|
|
if let Some(w) = phrase {
|
|
|
|
if !allowed_phrases.contains(w) {
|
|
|
|
new_phrase = Some(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
let mut new_split_words = None;
|
|
|
|
if let Some(w) = split_words {
|
|
|
|
if allowed_phrases.contains(w) {
|
|
|
|
new_split_words = Some(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
let mut new_synonyms = vec![];
|
|
|
|
for w in synonyms.iter() {
|
|
|
|
if allowed_phrases.contains(w) {
|
|
|
|
new_synonyms.push(*w);
|
|
|
|
} else {
|
|
|
|
changed = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if changed {
|
|
|
|
Some(QueryTerm {
|
|
|
|
original: *original,
|
|
|
|
is_ngram: *is_ngram,
|
|
|
|
is_prefix: *is_prefix,
|
|
|
|
phrase: new_phrase,
|
|
|
|
zero_typo: new_zero_typo,
|
|
|
|
prefix_of: new_prefix_of.into_boxed_slice(),
|
|
|
|
synonyms: new_synonyms.into_boxed_slice(),
|
|
|
|
split_words: new_split_words,
|
|
|
|
one_typo: new_one_typo.into_boxed_slice(),
|
|
|
|
two_typos: new_two_typos.into_boxed_slice(),
|
|
|
|
use_prefix_db: new_use_prefix_db,
|
|
|
|
})
|
|
|
|
} else {
|
|
|
|
None
|
|
|
|
}
|
|
|
|
}
|
2023-03-14 10:54:55 +01:00
|
|
|
pub fn phrase(
|
2023-03-14 16:37:47 +01:00
|
|
|
word_interner: &mut DedupInterner<String>,
|
|
|
|
phrase_interner: &mut DedupInterner<Phrase>,
|
2023-03-14 10:54:55 +01:00
|
|
|
phrase: Phrase,
|
|
|
|
) -> Self {
|
|
|
|
Self {
|
|
|
|
original: word_interner.insert(phrase.description(word_interner)),
|
|
|
|
phrase: Some(phrase_interner.insert(phrase)),
|
|
|
|
is_prefix: false,
|
|
|
|
zero_typo: None,
|
|
|
|
prefix_of: Box::new([]),
|
|
|
|
synonyms: Box::new([]),
|
|
|
|
split_words: None,
|
|
|
|
one_typo: Box::new([]),
|
|
|
|
two_typos: Box::new([]),
|
|
|
|
use_prefix_db: None,
|
|
|
|
is_ngram: false,
|
|
|
|
}
|
|
|
|
}
|
2023-03-14 16:37:47 +01:00
|
|
|
pub fn empty(word_interner: &mut DedupInterner<String>, original: &str) -> Self {
|
2023-03-13 17:21:29 +01:00
|
|
|
Self {
|
|
|
|
original: word_interner.insert(original.to_owned()),
|
2023-03-14 10:54:55 +01:00
|
|
|
phrase: None,
|
2023-03-13 17:21:29 +01:00
|
|
|
is_prefix: false,
|
|
|
|
zero_typo: None,
|
|
|
|
prefix_of: Box::new([]),
|
|
|
|
synonyms: Box::new([]),
|
|
|
|
split_words: None,
|
|
|
|
one_typo: Box::new([]),
|
|
|
|
two_typos: Box::new([]),
|
|
|
|
use_prefix_db: None,
|
2023-03-14 10:54:55 +01:00
|
|
|
is_ngram: false,
|
2023-03-13 17:21:29 +01:00
|
|
|
}
|
|
|
|
}
|
2023-03-08 15:04:25 +01:00
|
|
|
/// Return an iterator over all the single words derived from the original word.
|
|
|
|
///
|
|
|
|
/// This excludes synonyms, split words, and words stored in the prefix databases.
|
2023-03-14 10:54:55 +01:00
|
|
|
pub fn all_single_words_except_prefix_db(
|
2023-03-06 19:21:55 +01:00
|
|
|
&'_ self,
|
|
|
|
) -> impl Iterator<Item = Interned<String>> + Clone + '_ {
|
2023-03-13 17:21:29 +01:00
|
|
|
self.zero_typo
|
|
|
|
.iter()
|
|
|
|
.chain(self.prefix_of.iter())
|
|
|
|
.chain(self.one_typo.iter())
|
|
|
|
.chain(self.two_typos.iter())
|
|
|
|
.copied()
|
|
|
|
}
|
|
|
|
/// Return an iterator over all the single words derived from the original word.
|
|
|
|
///
|
|
|
|
/// This excludes synonyms, split words, and words stored in the prefix databases.
|
2023-03-14 10:54:55 +01:00
|
|
|
pub fn all_phrases(&'_ self) -> impl Iterator<Item = Interned<Phrase>> + Clone + '_ {
|
2023-03-13 17:21:29 +01:00
|
|
|
self.split_words.iter().chain(self.synonyms.iter()).copied()
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-09 11:12:31 +01:00
|
|
|
pub fn is_empty(&self) -> bool {
|
2023-03-13 17:21:29 +01:00
|
|
|
self.zero_typo.is_none()
|
2023-02-21 09:42:54 +01:00
|
|
|
&& self.one_typo.is_empty()
|
|
|
|
&& self.two_typos.is_empty()
|
2023-03-13 17:21:29 +01:00
|
|
|
&& self.prefix_of.is_empty()
|
2023-03-08 15:04:25 +01:00
|
|
|
&& self.synonyms.is_empty()
|
|
|
|
&& self.split_words.is_none()
|
2023-03-13 17:21:29 +01:00
|
|
|
&& self.use_prefix_db.is_none()
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-03-14 10:54:55 +01:00
|
|
|
/// Compute the query term for the given word
|
|
|
|
pub fn query_term_from_word(
|
2023-03-06 19:21:55 +01:00
|
|
|
ctx: &mut SearchContext,
|
2023-02-21 09:42:54 +01:00
|
|
|
word: &str,
|
|
|
|
max_typo: u8,
|
|
|
|
is_prefix: bool,
|
2023-03-14 10:54:55 +01:00
|
|
|
) -> Result<QueryTerm> {
|
2023-03-13 17:21:29 +01:00
|
|
|
if word.len() > MAX_WORD_LENGTH {
|
2023-03-14 10:54:55 +01:00
|
|
|
return Ok(QueryTerm::empty(&mut ctx.word_interner, word));
|
2023-03-13 17:21:29 +01:00
|
|
|
}
|
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
let fst = ctx.index.words_fst(ctx.txn)?;
|
2023-03-06 19:21:55 +01:00
|
|
|
let word_interned = ctx.word_interner.insert(word.to_owned());
|
|
|
|
|
2023-02-21 09:42:54 +01:00
|
|
|
let use_prefix_db = is_prefix
|
2023-03-06 19:21:55 +01:00
|
|
|
&& ctx
|
|
|
|
.index
|
|
|
|
.word_prefix_docids
|
|
|
|
.remap_data_type::<DecodeIgnore>()
|
|
|
|
.get(ctx.txn, word)?
|
|
|
|
.is_some();
|
2023-03-13 17:21:29 +01:00
|
|
|
let use_prefix_db = if use_prefix_db { Some(word_interned) } else { None };
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-13 17:21:29 +01:00
|
|
|
let mut zero_typo = None;
|
|
|
|
let mut prefix_of = vec![];
|
2023-02-21 09:42:54 +01:00
|
|
|
let mut one_typo = vec![];
|
|
|
|
let mut two_typos = vec![];
|
|
|
|
|
2023-03-13 17:21:29 +01:00
|
|
|
if fst.contains(word) {
|
|
|
|
zero_typo = Some(word_interned);
|
|
|
|
}
|
|
|
|
|
2023-02-21 09:42:54 +01:00
|
|
|
if max_typo == 0 {
|
2023-03-13 17:21:29 +01:00
|
|
|
if is_prefix && use_prefix_db.is_none() {
|
2023-02-21 09:42:54 +01:00
|
|
|
let prefix = Str::new(word).starts_with();
|
|
|
|
let mut stream = fst.search(prefix).into_stream();
|
|
|
|
|
2023-03-09 11:12:31 +01:00
|
|
|
while let Some(derived_word) = stream.next() {
|
|
|
|
let derived_word = std::str::from_utf8(derived_word)?.to_owned();
|
|
|
|
let derived_word_interned = ctx.word_interner.insert(derived_word);
|
2023-03-13 17:21:29 +01:00
|
|
|
if derived_word_interned != word_interned {
|
|
|
|
prefix_of.push(derived_word_interned);
|
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
} else if max_typo == 1 {
|
|
|
|
let dfa = build_dfa(word, 1, is_prefix);
|
|
|
|
let starts = StartsWith(Str::new(get_first(word)));
|
|
|
|
let mut stream = fst.search_with_state(Intersection(starts, &dfa)).into_stream();
|
2023-03-09 11:12:31 +01:00
|
|
|
// TODO: There may be wayyy too many matches (e.g. in the thousands), how to reduce them?
|
|
|
|
|
|
|
|
while let Some((derived_word, state)) = stream.next() {
|
|
|
|
let derived_word = std::str::from_utf8(derived_word)?;
|
2023-02-21 09:42:54 +01:00
|
|
|
|
|
|
|
let d = dfa.distance(state.1);
|
2023-03-09 11:12:31 +01:00
|
|
|
let derived_word_interned = ctx.word_interner.insert(derived_word.to_owned());
|
2023-02-21 09:42:54 +01:00
|
|
|
match d.to_u8() {
|
|
|
|
0 => {
|
2023-03-13 17:21:29 +01:00
|
|
|
if derived_word_interned != word_interned {
|
|
|
|
prefix_of.push(derived_word_interned);
|
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
1 => {
|
2023-03-09 11:12:31 +01:00
|
|
|
one_typo.push(derived_word_interned);
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
_ => panic!(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
let starts = StartsWith(Str::new(get_first(word)));
|
|
|
|
let first = Intersection(build_dfa(word, 1, is_prefix), Complement(&starts));
|
|
|
|
let second_dfa = build_dfa(word, 2, is_prefix);
|
|
|
|
let second = Intersection(&second_dfa, &starts);
|
|
|
|
let automaton = Union(first, &second);
|
|
|
|
|
|
|
|
let mut stream = fst.search_with_state(automaton).into_stream();
|
2023-03-09 11:12:31 +01:00
|
|
|
// TODO: There may be wayyy too many matches (e.g. in the thousands), how to reduce them?
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-09 11:12:31 +01:00
|
|
|
while let Some((derived_word, state)) = stream.next() {
|
|
|
|
let derived_word = std::str::from_utf8(derived_word)?;
|
|
|
|
let derived_word_interned = ctx.word_interner.insert(derived_word.to_owned());
|
2023-02-21 09:42:54 +01:00
|
|
|
// in the case the typo is on the first letter, we know the number of typo
|
|
|
|
// is two
|
2023-03-09 11:12:31 +01:00
|
|
|
if get_first(derived_word) != get_first(word) {
|
|
|
|
two_typos.push(derived_word_interned);
|
2023-02-21 09:42:54 +01:00
|
|
|
} else {
|
|
|
|
// Else, we know that it is the second dfa that matched and compute the
|
|
|
|
// correct distance
|
|
|
|
let d = second_dfa.distance((state.1).0);
|
|
|
|
match d.to_u8() {
|
|
|
|
0 => {
|
2023-03-13 17:21:29 +01:00
|
|
|
if derived_word_interned != word_interned {
|
|
|
|
prefix_of.push(derived_word_interned);
|
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
1 => {
|
2023-03-09 11:12:31 +01:00
|
|
|
one_typo.push(derived_word_interned);
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
2 => {
|
2023-03-09 11:12:31 +01:00
|
|
|
two_typos.push(derived_word_interned);
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
_ => panic!(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-03-06 19:21:55 +01:00
|
|
|
let split_words = split_best_frequency(ctx.index, ctx.txn, word)?.map(|(l, r)| {
|
|
|
|
ctx.phrase_interner.insert(Phrase {
|
|
|
|
words: vec![Some(ctx.word_interner.insert(l)), Some(ctx.word_interner.insert(r))],
|
|
|
|
})
|
|
|
|
});
|
|
|
|
|
|
|
|
let synonyms = ctx.index.synonyms(ctx.txn)?;
|
2023-03-02 21:27:57 +01:00
|
|
|
|
|
|
|
let synonyms = synonyms
|
|
|
|
.get(&vec![word.to_owned()])
|
|
|
|
.cloned()
|
|
|
|
.unwrap_or_default()
|
|
|
|
.into_iter()
|
2023-03-06 19:21:55 +01:00
|
|
|
.map(|words| {
|
|
|
|
let words = words.into_iter().map(|w| Some(ctx.word_interner.insert(w))).collect();
|
|
|
|
ctx.phrase_interner.insert(Phrase { words })
|
|
|
|
})
|
2023-03-02 21:27:57 +01:00
|
|
|
.collect();
|
|
|
|
|
2023-03-14 10:54:55 +01:00
|
|
|
Ok(QueryTerm {
|
2023-03-13 17:21:29 +01:00
|
|
|
original: word_interned,
|
2023-03-14 10:54:55 +01:00
|
|
|
phrase: None,
|
2023-03-13 17:21:29 +01:00
|
|
|
is_prefix,
|
|
|
|
zero_typo,
|
|
|
|
prefix_of: prefix_of.into_boxed_slice(),
|
2023-03-02 21:27:57 +01:00
|
|
|
synonyms,
|
|
|
|
split_words,
|
2023-03-06 19:21:55 +01:00
|
|
|
one_typo: one_typo.into_boxed_slice(),
|
|
|
|
two_typos: two_typos.into_boxed_slice(),
|
2023-03-02 21:27:57 +01:00
|
|
|
use_prefix_db,
|
2023-03-14 10:54:55 +01:00
|
|
|
is_ngram: false,
|
2023-03-13 17:21:29 +01:00
|
|
|
})
|
2023-03-02 21:27:57 +01:00
|
|
|
}
|
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
/// Split the original word into the two words that appear the
|
|
|
|
/// most next to each other in the index.
|
|
|
|
///
|
|
|
|
/// Return `None` if the original word cannot be split.
|
2023-03-02 21:27:57 +01:00
|
|
|
fn split_best_frequency(
|
|
|
|
index: &Index,
|
|
|
|
txn: &RoTxn,
|
|
|
|
original: &str,
|
|
|
|
) -> Result<Option<(String, String)>> {
|
|
|
|
let chars = original.char_indices().skip(1);
|
|
|
|
let mut best = None;
|
|
|
|
|
|
|
|
for (i, _) in chars {
|
|
|
|
let (left, right) = original.split_at(i);
|
|
|
|
|
|
|
|
let key = (1, left, right);
|
|
|
|
let frequency = index
|
|
|
|
.word_pair_proximity_docids
|
|
|
|
.remap_data_type::<CboRoaringBitmapLenCodec>()
|
|
|
|
.get(txn, &key)?
|
|
|
|
.unwrap_or(0);
|
|
|
|
|
|
|
|
if frequency != 0 && best.map_or(true, |(old, _, _)| frequency > old) {
|
|
|
|
best = Some((frequency, left, right));
|
|
|
|
}
|
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-02 21:27:57 +01:00
|
|
|
Ok(best.map(|(_, left, right)| (left.to_owned(), right.to_owned())))
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
impl QueryTerm {
|
2023-03-08 15:04:25 +01:00
|
|
|
/// Return the original word from the given query term
|
2023-03-14 10:54:55 +01:00
|
|
|
pub fn original_single_word(&self) -> Option<Interned<String>> {
|
|
|
|
if self.phrase.is_some() || self.is_ngram {
|
|
|
|
None
|
|
|
|
} else {
|
|
|
|
Some(self.original)
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
/// A query term term coupled with its position in the user's search query.
|
2023-03-06 19:21:55 +01:00
|
|
|
#[derive(Clone)]
|
2023-02-21 09:42:54 +01:00
|
|
|
pub struct LocatedQueryTerm {
|
2023-03-14 10:54:55 +01:00
|
|
|
pub value: Interned<QueryTerm>,
|
2023-02-21 09:42:54 +01:00
|
|
|
pub positions: RangeInclusive<i8>,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl LocatedQueryTerm {
|
2023-03-14 10:54:55 +01:00
|
|
|
/// Return `true` iff the term is empty
|
2023-03-14 16:37:47 +01:00
|
|
|
pub fn is_empty(&self, interner: &DedupInterner<QueryTerm>) -> bool {
|
2023-03-14 10:54:55 +01:00
|
|
|
interner.get(self.value).is_empty()
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
/// Convert the tokenised search query into a list of located query terms.
|
2023-03-23 09:15:57 +01:00
|
|
|
pub fn located_query_terms_from_string(
|
|
|
|
ctx: &mut SearchContext,
|
2023-03-23 09:35:53 +01:00
|
|
|
query: NormalizedTokenIter<&[u8]>,
|
2023-03-06 08:35:01 +01:00
|
|
|
words_limit: Option<usize>,
|
|
|
|
) -> Result<Vec<LocatedQueryTerm>> {
|
2023-03-13 17:21:29 +01:00
|
|
|
let nbr_typos = number_of_typos_allowed(ctx)?;
|
2023-03-06 08:35:01 +01:00
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
let mut located_terms = Vec::new();
|
2023-03-06 08:35:01 +01:00
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
let mut phrase = Vec::new();
|
2023-03-06 08:35:01 +01:00
|
|
|
let mut quoted = false;
|
|
|
|
|
|
|
|
let parts_limit = words_limit.unwrap_or(usize::MAX);
|
|
|
|
|
|
|
|
let mut position = -1i8;
|
|
|
|
let mut phrase_start = -1i8;
|
|
|
|
let mut phrase_end = -1i8;
|
|
|
|
|
|
|
|
let mut peekable = query.peekable();
|
|
|
|
while let Some(token) = peekable.next() {
|
|
|
|
// early return if word limit is exceeded
|
2023-03-08 15:04:25 +01:00
|
|
|
if located_terms.len() >= parts_limit {
|
|
|
|
return Ok(located_terms);
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
match token.kind {
|
|
|
|
TokenKind::Word | TokenKind::StopWord => {
|
|
|
|
position += 1;
|
|
|
|
// 1. if the word is quoted we push it in a phrase-buffer waiting for the ending quote,
|
|
|
|
// 2. if the word is not the last token of the query and is not a stop_word we push it as a non-prefix word,
|
|
|
|
// 3. if the word is the last token of the query we push it as a prefix word.
|
|
|
|
if quoted {
|
|
|
|
phrase_end = position;
|
|
|
|
if phrase.is_empty() {
|
|
|
|
phrase_start = position;
|
|
|
|
}
|
|
|
|
if let TokenKind::StopWord = token.kind {
|
|
|
|
phrase.push(None);
|
|
|
|
} else {
|
2023-03-06 19:21:55 +01:00
|
|
|
let word = ctx.word_interner.insert(token.lemma().to_string());
|
2023-03-06 08:35:01 +01:00
|
|
|
// TODO: in a phrase, check that every word exists
|
2023-03-14 10:54:55 +01:00
|
|
|
// otherwise return an empty term
|
2023-03-06 19:21:55 +01:00
|
|
|
phrase.push(Some(word));
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
|
|
|
} else if peekable.peek().is_some() {
|
|
|
|
match token.kind {
|
|
|
|
TokenKind::Word => {
|
2023-03-06 19:21:55 +01:00
|
|
|
let word = token.lemma();
|
2023-03-14 10:54:55 +01:00
|
|
|
let term = query_term_from_word(ctx, word, nbr_typos(word), false)?;
|
2023-02-21 09:42:54 +01:00
|
|
|
let located_term = LocatedQueryTerm {
|
2023-03-14 10:54:55 +01:00
|
|
|
value: ctx.term_interner.insert(term),
|
2023-02-21 09:42:54 +01:00
|
|
|
positions: position..=position,
|
|
|
|
};
|
2023-03-08 15:04:25 +01:00
|
|
|
located_terms.push(located_term);
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
TokenKind::StopWord | TokenKind::Separator(_) | TokenKind::Unknown => {}
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
} else {
|
2023-03-06 19:21:55 +01:00
|
|
|
let word = token.lemma();
|
2023-03-14 10:54:55 +01:00
|
|
|
let term = query_term_from_word(ctx, word, nbr_typos(word), true)?;
|
2023-03-06 08:35:01 +01:00
|
|
|
let located_term = LocatedQueryTerm {
|
2023-03-14 10:54:55 +01:00
|
|
|
value: ctx.term_interner.insert(term),
|
2023-03-06 08:35:01 +01:00
|
|
|
positions: position..=position,
|
|
|
|
};
|
2023-03-08 15:04:25 +01:00
|
|
|
located_terms.push(located_term);
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
|
|
|
TokenKind::Separator(separator_kind) => {
|
|
|
|
match separator_kind {
|
|
|
|
SeparatorKind::Hard => {
|
|
|
|
position += 1;
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
SeparatorKind::Soft => {
|
|
|
|
position += 0;
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
let quote_count = token.lemma().chars().filter(|&s| s == '"').count();
|
|
|
|
// swap quoted state if we encounter a double quote
|
|
|
|
if quote_count % 2 != 0 {
|
|
|
|
quoted = !quoted;
|
|
|
|
}
|
|
|
|
// if there is a quote or a hard separator we close the phrase.
|
|
|
|
if !phrase.is_empty() && (quote_count > 0 || separator_kind == SeparatorKind::Hard)
|
|
|
|
{
|
|
|
|
let located_query_term = LocatedQueryTerm {
|
2023-03-14 10:54:55 +01:00
|
|
|
value: ctx.term_interner.insert(QueryTerm::phrase(
|
|
|
|
&mut ctx.word_interner,
|
|
|
|
&mut ctx.phrase_interner,
|
|
|
|
Phrase { words: mem::take(&mut phrase) },
|
|
|
|
)),
|
2023-03-06 08:35:01 +01:00
|
|
|
positions: phrase_start..=phrase_end,
|
|
|
|
};
|
2023-03-08 15:04:25 +01:00
|
|
|
located_terms.push(located_query_term);
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
_ => (),
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-06 08:35:01 +01:00
|
|
|
// If a quote is never closed, we consider all of the end of the query as a phrase.
|
|
|
|
if !phrase.is_empty() {
|
|
|
|
let located_query_term = LocatedQueryTerm {
|
2023-03-14 10:54:55 +01:00
|
|
|
value: ctx.term_interner.insert(QueryTerm::phrase(
|
|
|
|
&mut ctx.word_interner,
|
|
|
|
&mut ctx.phrase_interner,
|
|
|
|
Phrase { words: mem::take(&mut phrase) },
|
|
|
|
)),
|
2023-03-06 08:35:01 +01:00
|
|
|
positions: phrase_start..=phrase_end,
|
|
|
|
};
|
2023-03-08 15:04:25 +01:00
|
|
|
located_terms.push(located_query_term);
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
|
2023-03-08 15:04:25 +01:00
|
|
|
Ok(located_terms)
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
|
2023-03-13 17:21:29 +01:00
|
|
|
pub fn number_of_typos_allowed<'ctx>(
|
|
|
|
ctx: &SearchContext<'ctx>,
|
|
|
|
) -> Result<impl Fn(&str) -> u8 + 'ctx> {
|
|
|
|
let authorize_typos = ctx.index.authorize_typos(ctx.txn)?;
|
|
|
|
let min_len_one_typo = ctx.index.min_word_len_one_typo(ctx.txn)?;
|
|
|
|
let min_len_two_typos = ctx.index.min_word_len_two_typos(ctx.txn)?;
|
|
|
|
|
|
|
|
// TODO: should `exact_words` also disable prefix search, ngrams, split words, or synonyms?
|
|
|
|
let exact_words = ctx.index.exact_words(ctx.txn)?;
|
|
|
|
|
|
|
|
Ok(Box::new(move |word: &str| {
|
|
|
|
if !authorize_typos
|
|
|
|
|| word.len() < min_len_one_typo as usize
|
|
|
|
|| exact_words.as_ref().map_or(false, |fst| fst.contains(word))
|
|
|
|
{
|
|
|
|
0
|
|
|
|
} else if word.len() < min_len_two_typos as usize {
|
|
|
|
1
|
|
|
|
} else {
|
|
|
|
2
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
2023-03-13 17:21:29 +01:00
|
|
|
}))
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
2023-03-08 15:04:25 +01:00
|
|
|
|
2023-03-13 17:21:29 +01:00
|
|
|
pub fn make_ngram(
|
2023-03-06 19:21:55 +01:00
|
|
|
ctx: &mut SearchContext,
|
2023-03-13 17:21:29 +01:00
|
|
|
terms: &[LocatedQueryTerm],
|
|
|
|
number_of_typos_allowed: &impl Fn(&str) -> u8,
|
|
|
|
) -> Result<Option<LocatedQueryTerm>> {
|
|
|
|
assert!(!terms.is_empty());
|
|
|
|
for ts in terms.windows(2) {
|
|
|
|
let [t1, t2] = ts else { panic!() };
|
|
|
|
if *t1.positions.end() != t2.positions.start() - 1 {
|
|
|
|
return Ok(None);
|
|
|
|
}
|
2023-03-06 08:35:01 +01:00
|
|
|
}
|
2023-03-13 17:21:29 +01:00
|
|
|
let mut words_interned = vec![];
|
|
|
|
for term in terms {
|
2023-03-14 10:54:55 +01:00
|
|
|
if let Some(original_term_word) = ctx.term_interner.get(term.value).original_single_word() {
|
2023-03-13 17:21:29 +01:00
|
|
|
words_interned.push(original_term_word);
|
|
|
|
} else {
|
|
|
|
return Ok(None);
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
}
|
2023-03-13 17:21:29 +01:00
|
|
|
let words =
|
|
|
|
words_interned.iter().map(|&i| ctx.word_interner.get(i).to_owned()).collect::<Vec<_>>();
|
|
|
|
|
|
|
|
let start = *terms.first().as_ref().unwrap().positions.start();
|
|
|
|
let end = *terms.last().as_ref().unwrap().positions.end();
|
2023-03-14 10:54:55 +01:00
|
|
|
let is_prefix = ctx.term_interner.get(terms.last().as_ref().unwrap().value).is_prefix;
|
2023-03-13 17:21:29 +01:00
|
|
|
let ngram_str = words.join("");
|
|
|
|
if ngram_str.len() > MAX_WORD_LENGTH {
|
|
|
|
return Ok(None);
|
|
|
|
}
|
|
|
|
|
2023-03-14 10:54:55 +01:00
|
|
|
let mut term = query_term_from_word(
|
2023-03-13 17:21:29 +01:00
|
|
|
ctx,
|
|
|
|
&ngram_str,
|
|
|
|
number_of_typos_allowed(ngram_str.as_str()).saturating_sub(terms.len() as u8),
|
|
|
|
is_prefix,
|
|
|
|
)?;
|
2023-03-14 10:54:55 +01:00
|
|
|
term.original = ctx.word_interner.insert(words.join(" "));
|
2023-03-13 17:21:29 +01:00
|
|
|
// Now add the synonyms
|
|
|
|
let index_synonyms = ctx.index.synonyms(ctx.txn)?;
|
2023-03-14 10:54:55 +01:00
|
|
|
let mut term_synonyms = term.synonyms.to_vec();
|
|
|
|
term_synonyms.extend(index_synonyms.get(&words).cloned().unwrap_or_default().into_iter().map(
|
|
|
|
|words| {
|
2023-03-13 17:21:29 +01:00
|
|
|
let words = words.into_iter().map(|w| Some(ctx.word_interner.insert(w))).collect();
|
|
|
|
ctx.phrase_interner.insert(Phrase { words })
|
2023-03-14 10:54:55 +01:00
|
|
|
},
|
|
|
|
));
|
|
|
|
term.synonyms = term_synonyms.into_boxed_slice();
|
|
|
|
if let Some(split_words) = term.split_words {
|
2023-03-13 17:21:29 +01:00
|
|
|
let split_words = ctx.phrase_interner.get(split_words);
|
|
|
|
if split_words.words == words_interned.iter().map(|&i| Some(i)).collect::<Vec<_>>() {
|
2023-03-14 10:54:55 +01:00
|
|
|
term.split_words = None;
|
2023-03-13 17:21:29 +01:00
|
|
|
}
|
|
|
|
}
|
2023-03-14 10:54:55 +01:00
|
|
|
if term.is_empty() {
|
2023-03-13 17:21:29 +01:00
|
|
|
return Ok(None);
|
|
|
|
}
|
2023-03-14 10:54:55 +01:00
|
|
|
term.is_ngram = true;
|
|
|
|
let term = LocatedQueryTerm { value: ctx.term_interner.insert(term), positions: start..=end };
|
2023-03-13 17:21:29 +01:00
|
|
|
|
|
|
|
Ok(Some(term))
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|