2023-02-21 09:42:54 +01:00
|
|
|
// TODO: put primitive query part in here
|
|
|
|
|
|
|
|
use std::borrow::Cow;
|
|
|
|
use std::mem;
|
|
|
|
use std::ops::RangeInclusive;
|
|
|
|
|
|
|
|
use charabia::normalizer::NormalizedTokenIter;
|
|
|
|
use charabia::{SeparatorKind, TokenKind};
|
|
|
|
use fst::automaton::Str;
|
|
|
|
use fst::{Automaton, IntoStreamer, Streamer};
|
|
|
|
use heed::types::DecodeIgnore;
|
|
|
|
use heed::RoTxn;
|
2023-03-02 21:27:57 +01:00
|
|
|
use itertools::Itertools;
|
2023-02-21 09:42:54 +01:00
|
|
|
|
|
|
|
use crate::search::fst_utils::{Complement, Intersection, StartsWith, Union};
|
|
|
|
use crate::search::{build_dfa, get_first};
|
2023-03-02 21:27:57 +01:00
|
|
|
use crate::{CboRoaringBitmapLenCodec, Index, Result};
|
|
|
|
|
|
|
|
#[derive(Debug, Default, Clone)]
|
|
|
|
pub struct Phrase {
|
|
|
|
pub words: Vec<Option<String>>,
|
|
|
|
}
|
|
|
|
impl Phrase {
|
|
|
|
pub fn description(&self) -> String {
|
|
|
|
self.words.iter().flatten().join(" ")
|
|
|
|
}
|
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
|
|
|
|
#[derive(Debug, Clone)]
|
|
|
|
pub struct WordDerivations {
|
|
|
|
pub original: String,
|
2023-03-02 21:27:57 +01:00
|
|
|
// TODO: pub prefix_of: Vec<String>,
|
|
|
|
pub synonyms: Vec<Phrase>,
|
|
|
|
pub split_words: Option<(String, String)>,
|
2023-02-21 09:42:54 +01:00
|
|
|
pub zero_typo: Vec<String>,
|
|
|
|
pub one_typo: Vec<String>,
|
|
|
|
pub two_typos: Vec<String>,
|
|
|
|
pub use_prefix_db: bool,
|
|
|
|
}
|
|
|
|
impl WordDerivations {
|
|
|
|
pub fn all_derivations_except_prefix_db(&self) -> impl Iterator<Item = &String> + Clone {
|
|
|
|
self.zero_typo.iter().chain(self.one_typo.iter()).chain(self.two_typos.iter())
|
|
|
|
}
|
|
|
|
fn is_empty(&self) -> bool {
|
|
|
|
self.zero_typo.is_empty()
|
|
|
|
&& self.one_typo.is_empty()
|
|
|
|
&& self.two_typos.is_empty()
|
|
|
|
&& !self.use_prefix_db
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn word_derivations(
|
|
|
|
index: &Index,
|
|
|
|
txn: &RoTxn,
|
|
|
|
word: &str,
|
|
|
|
max_typo: u8,
|
|
|
|
is_prefix: bool,
|
|
|
|
fst: &fst::Set<Cow<[u8]>>,
|
|
|
|
) -> Result<WordDerivations> {
|
|
|
|
let use_prefix_db = is_prefix
|
|
|
|
&& index.word_prefix_docids.remap_data_type::<DecodeIgnore>().get(txn, word)?.is_some();
|
|
|
|
|
|
|
|
let mut zero_typo = vec![];
|
|
|
|
let mut one_typo = vec![];
|
|
|
|
let mut two_typos = vec![];
|
|
|
|
|
|
|
|
if max_typo == 0 {
|
2023-02-23 13:13:19 +01:00
|
|
|
if is_prefix && !use_prefix_db {
|
2023-02-21 09:42:54 +01:00
|
|
|
let prefix = Str::new(word).starts_with();
|
|
|
|
let mut stream = fst.search(prefix).into_stream();
|
|
|
|
|
|
|
|
while let Some(word) = stream.next() {
|
|
|
|
let word = std::str::from_utf8(word)?;
|
|
|
|
zero_typo.push(word.to_string());
|
|
|
|
}
|
|
|
|
} else if fst.contains(word) {
|
|
|
|
zero_typo.push(word.to_string());
|
|
|
|
}
|
|
|
|
} else if max_typo == 1 {
|
|
|
|
let dfa = build_dfa(word, 1, is_prefix);
|
|
|
|
let starts = StartsWith(Str::new(get_first(word)));
|
|
|
|
let mut stream = fst.search_with_state(Intersection(starts, &dfa)).into_stream();
|
|
|
|
|
|
|
|
while let Some((word, state)) = stream.next() {
|
|
|
|
let word = std::str::from_utf8(word)?;
|
|
|
|
let d = dfa.distance(state.1);
|
|
|
|
match d.to_u8() {
|
|
|
|
0 => {
|
|
|
|
zero_typo.push(word.to_string());
|
|
|
|
}
|
|
|
|
1 => {
|
|
|
|
one_typo.push(word.to_string());
|
|
|
|
}
|
|
|
|
_ => panic!(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
let starts = StartsWith(Str::new(get_first(word)));
|
|
|
|
let first = Intersection(build_dfa(word, 1, is_prefix), Complement(&starts));
|
|
|
|
let second_dfa = build_dfa(word, 2, is_prefix);
|
|
|
|
let second = Intersection(&second_dfa, &starts);
|
|
|
|
let automaton = Union(first, &second);
|
|
|
|
|
|
|
|
let mut stream = fst.search_with_state(automaton).into_stream();
|
|
|
|
|
|
|
|
while let Some((found_word, state)) = stream.next() {
|
|
|
|
let found_word = std::str::from_utf8(found_word)?;
|
|
|
|
// in the case the typo is on the first letter, we know the number of typo
|
|
|
|
// is two
|
|
|
|
if get_first(found_word) != get_first(word) {
|
|
|
|
two_typos.push(found_word.to_string());
|
|
|
|
} else {
|
|
|
|
// Else, we know that it is the second dfa that matched and compute the
|
|
|
|
// correct distance
|
|
|
|
let d = second_dfa.distance((state.1).0);
|
|
|
|
match d.to_u8() {
|
|
|
|
0 => {
|
|
|
|
zero_typo.push(found_word.to_string());
|
|
|
|
}
|
|
|
|
1 => {
|
|
|
|
one_typo.push(found_word.to_string());
|
|
|
|
}
|
|
|
|
2 => {
|
|
|
|
two_typos.push(found_word.to_string());
|
|
|
|
}
|
|
|
|
_ => panic!(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-03-02 21:27:57 +01:00
|
|
|
let split_words = split_best_frequency(index, txn, word)?;
|
|
|
|
|
|
|
|
let synonyms = index.synonyms(txn)?;
|
|
|
|
let synonyms = synonyms
|
|
|
|
.get(&vec![word.to_owned()])
|
|
|
|
.cloned()
|
|
|
|
.unwrap_or_default()
|
|
|
|
.into_iter()
|
|
|
|
.map(|words| Phrase { words: words.into_iter().map(Some).collect() })
|
|
|
|
.collect();
|
|
|
|
|
|
|
|
Ok(WordDerivations {
|
|
|
|
original: word.to_owned(),
|
|
|
|
synonyms,
|
|
|
|
split_words,
|
|
|
|
zero_typo,
|
|
|
|
one_typo,
|
|
|
|
two_typos,
|
|
|
|
use_prefix_db,
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
fn split_best_frequency(
|
|
|
|
index: &Index,
|
|
|
|
txn: &RoTxn,
|
|
|
|
original: &str,
|
|
|
|
) -> Result<Option<(String, String)>> {
|
|
|
|
let chars = original.char_indices().skip(1);
|
|
|
|
let mut best = None;
|
|
|
|
|
|
|
|
for (i, _) in chars {
|
|
|
|
let (left, right) = original.split_at(i);
|
|
|
|
|
|
|
|
let key = (1, left, right);
|
|
|
|
let frequency = index
|
|
|
|
.word_pair_proximity_docids
|
|
|
|
.remap_data_type::<CboRoaringBitmapLenCodec>()
|
|
|
|
.get(txn, &key)?
|
|
|
|
.unwrap_or(0);
|
|
|
|
|
|
|
|
if frequency != 0 && best.map_or(true, |(old, _, _)| frequency > old) {
|
|
|
|
best = Some((frequency, left, right));
|
|
|
|
}
|
|
|
|
}
|
2023-02-21 09:42:54 +01:00
|
|
|
|
2023-03-02 21:27:57 +01:00
|
|
|
Ok(best.map(|(_, left, right)| (left.to_owned(), right.to_owned())))
|
2023-02-21 09:42:54 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Debug, Clone)]
|
|
|
|
pub enum QueryTerm {
|
2023-03-02 21:27:57 +01:00
|
|
|
Phrase { phrase: Phrase },
|
2023-02-21 09:42:54 +01:00
|
|
|
Word { derivations: WordDerivations },
|
|
|
|
}
|
|
|
|
impl QueryTerm {
|
|
|
|
pub fn original_single_word(&self) -> Option<&str> {
|
|
|
|
match self {
|
2023-03-02 21:27:57 +01:00
|
|
|
QueryTerm::Phrase { phrase: _ } => None,
|
2023-02-21 09:42:54 +01:00
|
|
|
QueryTerm::Word { derivations } => {
|
|
|
|
if derivations.is_empty() {
|
|
|
|
None
|
|
|
|
} else {
|
|
|
|
Some(derivations.original.as_str())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[derive(Debug, Clone)]
|
|
|
|
pub struct LocatedQueryTerm {
|
2023-03-02 21:27:57 +01:00
|
|
|
pub value: QueryTerm,
|
2023-02-21 09:42:54 +01:00
|
|
|
pub positions: RangeInclusive<i8>,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl LocatedQueryTerm {
|
|
|
|
pub fn is_empty(&self) -> bool {
|
|
|
|
match &self.value {
|
2023-03-02 21:27:57 +01:00
|
|
|
QueryTerm::Phrase { phrase: _ } => false,
|
2023-02-21 09:42:54 +01:00
|
|
|
QueryTerm::Word { derivations, .. } => derivations.is_empty(),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/// Create primitive query from tokenized query string,
|
|
|
|
/// the primitive query is an intermediate state to build the query tree.
|
|
|
|
pub fn from_query(
|
|
|
|
query: NormalizedTokenIter<Vec<u8>>,
|
|
|
|
words_limit: Option<usize>,
|
2023-03-02 21:27:57 +01:00
|
|
|
// TODO:` use index + txn + ? instead of closure
|
2023-02-21 09:42:54 +01:00
|
|
|
derivations: impl Fn(&str, bool) -> Result<WordDerivations>,
|
|
|
|
) -> Result<Vec<LocatedQueryTerm>> {
|
|
|
|
let mut primitive_query = Vec::new();
|
|
|
|
let mut phrase = Vec::new();
|
|
|
|
|
|
|
|
let mut quoted = false;
|
|
|
|
|
|
|
|
let parts_limit = words_limit.unwrap_or(usize::MAX);
|
|
|
|
|
|
|
|
let mut position = -1i8;
|
|
|
|
let mut phrase_start = -1i8;
|
|
|
|
let mut phrase_end = -1i8;
|
|
|
|
|
|
|
|
let mut peekable = query.peekable();
|
|
|
|
while let Some(token) = peekable.next() {
|
|
|
|
// early return if word limit is exceeded
|
|
|
|
if primitive_query.len() >= parts_limit {
|
|
|
|
return Ok(primitive_query);
|
|
|
|
}
|
|
|
|
|
|
|
|
match token.kind {
|
|
|
|
TokenKind::Word | TokenKind::StopWord => {
|
|
|
|
position += 1;
|
|
|
|
// 1. if the word is quoted we push it in a phrase-buffer waiting for the ending quote,
|
|
|
|
// 2. if the word is not the last token of the query and is not a stop_word we push it as a non-prefix word,
|
|
|
|
// 3. if the word is the last token of the query we push it as a prefix word.
|
|
|
|
if quoted {
|
|
|
|
phrase_end = position;
|
|
|
|
if phrase.is_empty() {
|
|
|
|
phrase_start = position;
|
|
|
|
}
|
|
|
|
if let TokenKind::StopWord = token.kind {
|
|
|
|
phrase.push(None);
|
|
|
|
} else {
|
|
|
|
// TODO: in a phrase, check that every word exists
|
|
|
|
// otherwise return WordDerivations::Empty
|
|
|
|
phrase.push(Some(token.lemma().to_string()));
|
|
|
|
}
|
|
|
|
} else if peekable.peek().is_some() {
|
|
|
|
if let TokenKind::StopWord = token.kind {
|
|
|
|
} else {
|
|
|
|
let derivations = derivations(token.lemma(), false)?;
|
|
|
|
let located_term = LocatedQueryTerm {
|
|
|
|
value: QueryTerm::Word { derivations },
|
|
|
|
positions: position..=position,
|
|
|
|
};
|
|
|
|
primitive_query.push(located_term);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
let derivations = derivations(token.lemma(), true)?;
|
|
|
|
let located_term = LocatedQueryTerm {
|
|
|
|
value: QueryTerm::Word { derivations },
|
|
|
|
positions: position..=position,
|
|
|
|
};
|
|
|
|
primitive_query.push(located_term);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
TokenKind::Separator(separator_kind) => {
|
|
|
|
match separator_kind {
|
|
|
|
SeparatorKind::Hard => {
|
|
|
|
position += 1;
|
|
|
|
}
|
|
|
|
SeparatorKind::Soft => {
|
|
|
|
position += 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
let quote_count = token.lemma().chars().filter(|&s| s == '"').count();
|
|
|
|
// swap quoted state if we encounter a double quote
|
|
|
|
if quote_count % 2 != 0 {
|
|
|
|
quoted = !quoted;
|
|
|
|
}
|
|
|
|
// if there is a quote or a hard separator we close the phrase.
|
|
|
|
if !phrase.is_empty()
|
|
|
|
&& (quote_count > 0 || separator_kind == SeparatorKind::Hard)
|
|
|
|
{
|
|
|
|
let located_query_term = LocatedQueryTerm {
|
2023-03-02 21:27:57 +01:00
|
|
|
value: QueryTerm::Phrase {
|
|
|
|
phrase: Phrase { words: mem::take(&mut phrase) },
|
|
|
|
},
|
2023-02-21 09:42:54 +01:00
|
|
|
positions: phrase_start..=phrase_end,
|
|
|
|
};
|
|
|
|
primitive_query.push(located_query_term);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
_ => (),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// If a quote is never closed, we consider all of the end of the query as a phrase.
|
|
|
|
if !phrase.is_empty() {
|
|
|
|
let located_query_term = LocatedQueryTerm {
|
2023-03-02 21:27:57 +01:00
|
|
|
value: QueryTerm::Phrase { phrase: Phrase { words: mem::take(&mut phrase) } },
|
2023-02-21 09:42:54 +01:00
|
|
|
positions: phrase_start..=phrase_end,
|
|
|
|
};
|
|
|
|
primitive_query.push(located_query_term);
|
|
|
|
}
|
|
|
|
|
|
|
|
Ok(primitive_query)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
impl LocatedQueryTerm {
|
|
|
|
pub fn ngram2(
|
|
|
|
x: &LocatedQueryTerm,
|
|
|
|
y: &LocatedQueryTerm,
|
|
|
|
) -> Option<(String, RangeInclusive<i8>)> {
|
|
|
|
if *x.positions.end() != y.positions.start() - 1 {
|
|
|
|
println!(
|
|
|
|
"x positions end: {}, y positions start: {}",
|
|
|
|
*x.positions.end(),
|
|
|
|
y.positions.start()
|
|
|
|
);
|
|
|
|
return None;
|
|
|
|
}
|
|
|
|
match (&x.value.original_single_word(), &y.value.original_single_word()) {
|
|
|
|
(Some(w1), Some(w2)) => {
|
|
|
|
let term = (format!("{w1}{w2}"), *x.positions.start()..=*y.positions.end());
|
|
|
|
Some(term)
|
|
|
|
}
|
|
|
|
_ => None,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
pub fn ngram3(
|
|
|
|
x: &LocatedQueryTerm,
|
|
|
|
y: &LocatedQueryTerm,
|
|
|
|
z: &LocatedQueryTerm,
|
|
|
|
) -> Option<(String, RangeInclusive<i8>)> {
|
|
|
|
if *x.positions.end() != y.positions.start() - 1
|
|
|
|
|| *y.positions.end() != z.positions.start() - 1
|
|
|
|
{
|
|
|
|
return None;
|
|
|
|
}
|
|
|
|
match (
|
|
|
|
&x.value.original_single_word(),
|
|
|
|
&y.value.original_single_word(),
|
|
|
|
&z.value.original_single_word(),
|
|
|
|
) {
|
|
|
|
(Some(w1), Some(w2), Some(w3)) => {
|
|
|
|
let term = (format!("{w1}{w2}{w3}"), *x.positions.start()..=*z.positions.end());
|
|
|
|
Some(term)
|
|
|
|
}
|
|
|
|
_ => None,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|