mirror of
https://github.com/meilisearch/MeiliSearch
synced 2024-11-26 06:44:27 +01:00
Switch string facet levels indexation to new algo
Write the algorithm once for both numbers and strings
This commit is contained in:
parent
cf0cd92ed4
commit
8d4b21a005
@ -1,13 +1,12 @@
|
|||||||
use std::fs::File;
|
|
||||||
use std::num::{NonZeroU8, NonZeroUsize};
|
|
||||||
use std::ops::RangeInclusive;
|
|
||||||
use std::{cmp, mem};
|
|
||||||
|
|
||||||
use grenad::{CompressionType, Reader, Writer};
|
use grenad::{CompressionType, Reader, Writer};
|
||||||
use heed::types::{ByteSlice, DecodeIgnore};
|
use heed::types::{ByteSlice, DecodeIgnore};
|
||||||
use heed::{BytesDecode, BytesEncode, Error};
|
use heed::{BytesDecode, BytesEncode, Error};
|
||||||
use log::debug;
|
use log::debug;
|
||||||
use roaring::RoaringBitmap;
|
use roaring::RoaringBitmap;
|
||||||
|
use std::cmp;
|
||||||
|
use std::fs::File;
|
||||||
|
use std::num::{NonZeroU8, NonZeroUsize};
|
||||||
|
use std::ops::RangeFrom;
|
||||||
use time::OffsetDateTime;
|
use time::OffsetDateTime;
|
||||||
|
|
||||||
use crate::error::InternalError;
|
use crate::error::InternalError;
|
||||||
@ -66,14 +65,7 @@ impl<'t, 'u, 'i> Facets<'t, 'u, 'i> {
|
|||||||
field_id,
|
field_id,
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
// Compute and store the faceted strings documents ids.
|
let (facet_string_levels, string_documents_ids) = compute_facet_strings_levels(
|
||||||
let string_documents_ids = compute_faceted_strings_documents_ids(
|
|
||||||
self.wtxn,
|
|
||||||
self.index.facet_id_string_docids.remap_key_type::<ByteSlice>(),
|
|
||||||
field_id,
|
|
||||||
)?;
|
|
||||||
|
|
||||||
let facet_string_levels = compute_facet_string_levels(
|
|
||||||
self.wtxn,
|
self.wtxn,
|
||||||
self.index.facet_id_string_docids,
|
self.index.facet_id_string_docids,
|
||||||
self.chunk_compression_type,
|
self.chunk_compression_type,
|
||||||
@ -83,36 +75,26 @@ impl<'t, 'u, 'i> Facets<'t, 'u, 'i> {
|
|||||||
field_id,
|
field_id,
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
|
self.index.put_string_faceted_documents_ids(
|
||||||
|
self.wtxn,
|
||||||
|
field_id,
|
||||||
|
&string_documents_ids,
|
||||||
|
)?;
|
||||||
|
for facet_strings_levels in facet_string_levels {
|
||||||
|
write_into_lmdb_database(
|
||||||
|
self.wtxn,
|
||||||
|
*self.index.facet_id_string_docids.as_polymorph(),
|
||||||
|
facet_strings_levels,
|
||||||
|
|_, _| {
|
||||||
|
Err(InternalError::IndexingMergingKeys { process: "facet string levels" })?
|
||||||
|
},
|
||||||
|
)?;
|
||||||
|
}
|
||||||
|
|
||||||
// Clear the facet number levels.
|
// Clear the facet number levels.
|
||||||
clear_field_number_levels(self.wtxn, self.index.facet_id_f64_docids, field_id)?;
|
clear_field_number_levels(self.wtxn, self.index.facet_id_f64_docids, field_id)?;
|
||||||
|
|
||||||
// Compute and store the faceted numbers documents ids.
|
let (facet_number_levels_2, number_documents_ids) = compute_facet_number_levels(
|
||||||
// let number_documents_ids = compute_faceted_numbers_documents_ids(
|
|
||||||
// self.wtxn,
|
|
||||||
// self.index.facet_id_f64_docids.remap_key_type::<ByteSlice>(),
|
|
||||||
// field_id,
|
|
||||||
// )?;
|
|
||||||
|
|
||||||
// let facet_number_levels = compute_facet_number_levels(
|
|
||||||
// self.wtxn,
|
|
||||||
// self.index.facet_id_f64_docids,
|
|
||||||
// self.chunk_compression_type,
|
|
||||||
// self.chunk_compression_level,
|
|
||||||
// self.level_group_size,
|
|
||||||
// self.min_level_size,
|
|
||||||
// field_id,
|
|
||||||
// )?;
|
|
||||||
|
|
||||||
// println!("printing 1");
|
|
||||||
|
|
||||||
// let mut cursor = facet_number_levels.into_cursor().unwrap();
|
|
||||||
// while let Some((key, bitmap)) = cursor.move_on_next().unwrap() {
|
|
||||||
// let key = FacetLevelValueF64Codec::bytes_decode(key).unwrap();
|
|
||||||
// let bitmap = CboRoaringBitmapCodec::bytes_decode(bitmap).unwrap();
|
|
||||||
// println!("{key:?} {bitmap:?}");
|
|
||||||
// }
|
|
||||||
|
|
||||||
let (facet_number_levels_2, number_documents_ids) = compute_facet_number_levels_2(
|
|
||||||
self.wtxn,
|
self.wtxn,
|
||||||
self.index.facet_id_f64_docids,
|
self.index.facet_id_f64_docids,
|
||||||
self.chunk_compression_type,
|
self.chunk_compression_type,
|
||||||
@ -122,37 +104,6 @@ impl<'t, 'u, 'i> Facets<'t, 'u, 'i> {
|
|||||||
field_id,
|
field_id,
|
||||||
)?;
|
)?;
|
||||||
|
|
||||||
// let mut writer = create_writer(
|
|
||||||
// self.chunk_compression_type,
|
|
||||||
// self.chunk_compression_level,
|
|
||||||
// tempfile::tempfile()?,
|
|
||||||
// );
|
|
||||||
// for fnl in facet_number_levels_2 {
|
|
||||||
// let mut cursor = fnl.into_cursor().unwrap();
|
|
||||||
// while let Some((key, bitmap)) = cursor.move_on_next().unwrap() {
|
|
||||||
// writer.insert(key, bitmap).unwrap();
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// let reader = writer_into_reader(writer)?;
|
|
||||||
// let mut cursor1 = reader.into_cursor().unwrap();
|
|
||||||
// let mut cursor2 = facet_number_levels.into_cursor().unwrap();
|
|
||||||
// loop {
|
|
||||||
// let (c1, c2) = (cursor1.move_on_next().unwrap(), cursor2.move_on_next().unwrap());
|
|
||||||
// match (c1, c2) {
|
|
||||||
// (Some((k1, v1)), Some((k2, v2))) => {
|
|
||||||
// assert_eq!(k1, k2);
|
|
||||||
// assert_eq!(v1, v2);
|
|
||||||
// }
|
|
||||||
// (None, None) => break,
|
|
||||||
// _ => panic!(),
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
||||||
self.index.put_string_faceted_documents_ids(
|
|
||||||
self.wtxn,
|
|
||||||
field_id,
|
|
||||||
&string_documents_ids,
|
|
||||||
)?;
|
|
||||||
self.index.put_number_faceted_documents_ids(
|
self.index.put_number_faceted_documents_ids(
|
||||||
self.wtxn,
|
self.wtxn,
|
||||||
field_id,
|
field_id,
|
||||||
@ -169,31 +120,13 @@ impl<'t, 'u, 'i> Facets<'t, 'u, 'i> {
|
|||||||
},
|
},
|
||||||
)?;
|
)?;
|
||||||
}
|
}
|
||||||
|
|
||||||
write_into_lmdb_database(
|
|
||||||
self.wtxn,
|
|
||||||
*self.index.facet_id_string_docids.as_polymorph(),
|
|
||||||
facet_string_levels,
|
|
||||||
|_, _| Err(InternalError::IndexingMergingKeys { process: "facet string levels" })?,
|
|
||||||
)?;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn clear_field_number_levels<'t>(
|
fn compute_facet_number_levels<'t>(
|
||||||
wtxn: &'t mut heed::RwTxn,
|
|
||||||
db: heed::Database<FacetLevelValueF64Codec, CboRoaringBitmapCodec>,
|
|
||||||
field_id: FieldId,
|
|
||||||
) -> heed::Result<()> {
|
|
||||||
let left = (field_id, 1, f64::MIN, f64::MIN);
|
|
||||||
let right = (field_id, u8::MAX, f64::MAX, f64::MAX);
|
|
||||||
let range = left..=right;
|
|
||||||
db.delete_range(wtxn, &range).map(drop)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn compute_facet_number_levels_2<'t>(
|
|
||||||
rtxn: &'t heed::RoTxn,
|
rtxn: &'t heed::RoTxn,
|
||||||
db: heed::Database<FacetLevelValueF64Codec, CboRoaringBitmapCodec>,
|
db: heed::Database<FacetLevelValueF64Codec, CboRoaringBitmapCodec>,
|
||||||
compression_type: CompressionType,
|
compression_type: CompressionType,
|
||||||
@ -208,11 +141,7 @@ fn compute_facet_number_levels_2<'t>(
|
|||||||
.remap_types::<DecodeIgnore, DecodeIgnore>()
|
.remap_types::<DecodeIgnore, DecodeIgnore>()
|
||||||
.fold(Ok(0usize), |count, result| result.and(count).map(|c| c + 1))?;
|
.fold(Ok(0usize), |count, result| result.and(count).map(|c| c + 1))?;
|
||||||
|
|
||||||
let level_0_range = {
|
let level_0_start = (field_id, 0, f64::MIN, f64::MIN);
|
||||||
let left = (field_id, 0, f64::MIN, f64::MIN);
|
|
||||||
let right = (field_id, 0, f64::MAX, f64::MAX);
|
|
||||||
left..=right
|
|
||||||
};
|
|
||||||
|
|
||||||
// Groups sizes are always a power of the original level_group_size and therefore a group
|
// Groups sizes are always a power of the original level_group_size and therefore a group
|
||||||
// always maps groups of the previous level and never splits previous levels groups in half.
|
// always maps groups of the previous level and never splits previous levels groups in half.
|
||||||
@ -221,32 +150,38 @@ fn compute_facet_number_levels_2<'t>(
|
|||||||
.take_while(|(_, s)| first_level_size / *s >= min_level_size.get())
|
.take_while(|(_, s)| first_level_size / *s >= min_level_size.get())
|
||||||
.collect::<Vec<_>>();
|
.collect::<Vec<_>>();
|
||||||
|
|
||||||
// dbg!(first_level_size, min_level_size);
|
|
||||||
// dbg!(level_group_size);
|
|
||||||
// dbg!(&group_size_iter);
|
|
||||||
|
|
||||||
let mut number_document_ids = RoaringBitmap::new();
|
let mut number_document_ids = RoaringBitmap::new();
|
||||||
|
|
||||||
if let Some((top_level, _)) = group_size_iter.last() {
|
if let Some((top_level, _)) = group_size_iter.last() {
|
||||||
let subwriters = recursive_compute_levels(
|
let subwriters =
|
||||||
rtxn,
|
recursive_compute_levels::<FacetLevelValueF64Codec, CboRoaringBitmapCodec, f64>(
|
||||||
db,
|
rtxn,
|
||||||
compression_type,
|
db,
|
||||||
compression_level,
|
compression_type,
|
||||||
*top_level,
|
compression_level,
|
||||||
level_0_range,
|
*top_level,
|
||||||
level_group_size,
|
level_0_start,
|
||||||
&mut |bitmaps, _, _| {
|
&(level_0_start..),
|
||||||
for bitmap in bitmaps {
|
first_level_size,
|
||||||
number_document_ids |= bitmap;
|
level_group_size,
|
||||||
}
|
&mut |bitmaps, _, _| {
|
||||||
Ok(())
|
for bitmap in bitmaps {
|
||||||
},
|
number_document_ids |= bitmap;
|
||||||
)?;
|
}
|
||||||
|
Ok(())
|
||||||
|
},
|
||||||
|
&|_i, (_field_id, _level, left, _right)| *left,
|
||||||
|
&|bitmap| bitmap,
|
||||||
|
&|writer, level, left, right, docids| {
|
||||||
|
write_number_entry(writer, field_id, level.get(), left, right, &docids)?;
|
||||||
|
Ok(())
|
||||||
|
},
|
||||||
|
)?;
|
||||||
|
|
||||||
Ok((subwriters, number_document_ids))
|
Ok((subwriters, number_document_ids))
|
||||||
} else {
|
} else {
|
||||||
let mut documents_ids = RoaringBitmap::new();
|
let mut documents_ids = RoaringBitmap::new();
|
||||||
for result in db.range(rtxn, &level_0_range)? {
|
for result in db.range(rtxn, &(level_0_start..))?.take(first_level_size) {
|
||||||
let (_key, docids) = result?;
|
let (_key, docids) = result?;
|
||||||
documents_ids |= docids;
|
documents_ids |= docids;
|
||||||
}
|
}
|
||||||
@ -255,52 +190,129 @@ fn compute_facet_number_levels_2<'t>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn recursive_compute_levels<'t>(
|
fn compute_facet_strings_levels<'t>(
|
||||||
rtxn: &'t heed::RoTxn,
|
rtxn: &'t heed::RoTxn,
|
||||||
db: heed::Database<FacetLevelValueF64Codec, CboRoaringBitmapCodec>,
|
db: heed::Database<FacetStringLevelZeroCodec, FacetStringLevelZeroValueCodec>,
|
||||||
|
compression_type: CompressionType,
|
||||||
|
compression_level: Option<u32>,
|
||||||
|
level_group_size: NonZeroUsize,
|
||||||
|
min_level_size: NonZeroUsize,
|
||||||
|
field_id: FieldId,
|
||||||
|
) -> Result<(Vec<Reader<File>>, RoaringBitmap)> {
|
||||||
|
let first_level_size = db
|
||||||
|
.remap_key_type::<ByteSlice>()
|
||||||
|
.prefix_iter(rtxn, &field_id.to_be_bytes())?
|
||||||
|
.remap_types::<DecodeIgnore, DecodeIgnore>()
|
||||||
|
.fold(Ok(0usize), |count, result| result.and(count).map(|c| c + 1))?;
|
||||||
|
|
||||||
|
let level_0_start = (field_id, "");
|
||||||
|
|
||||||
|
// Groups sizes are always a power of the original level_group_size and therefore a group
|
||||||
|
// always maps groups of the previous level and never splits previous levels groups in half.
|
||||||
|
let group_size_iter = (1u8..)
|
||||||
|
.map(|l| (l, level_group_size.get().pow(l as u32)))
|
||||||
|
.take_while(|(_, s)| first_level_size / *s >= min_level_size.get())
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
|
||||||
|
let mut strings_document_ids = RoaringBitmap::new();
|
||||||
|
|
||||||
|
if let Some((top_level, _)) = group_size_iter.last() {
|
||||||
|
let subwriters = recursive_compute_levels::<
|
||||||
|
FacetStringLevelZeroCodec,
|
||||||
|
FacetStringLevelZeroValueCodec,
|
||||||
|
(u32, &str),
|
||||||
|
>(
|
||||||
|
rtxn,
|
||||||
|
db,
|
||||||
|
compression_type,
|
||||||
|
compression_level,
|
||||||
|
*top_level,
|
||||||
|
level_0_start,
|
||||||
|
&(level_0_start..),
|
||||||
|
first_level_size,
|
||||||
|
level_group_size,
|
||||||
|
&mut |bitmaps, _, _| {
|
||||||
|
for bitmap in bitmaps {
|
||||||
|
strings_document_ids |= bitmap;
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
},
|
||||||
|
&|i, (_field_id, value)| (i as u32, *value),
|
||||||
|
&|value| value.1,
|
||||||
|
&|writer, level, start_bound, end_bound, docids| {
|
||||||
|
write_string_entry(writer, field_id, level, start_bound, end_bound, docids)?;
|
||||||
|
Ok(())
|
||||||
|
},
|
||||||
|
)?;
|
||||||
|
|
||||||
|
Ok((subwriters, strings_document_ids))
|
||||||
|
} else {
|
||||||
|
let mut documents_ids = RoaringBitmap::new();
|
||||||
|
for result in db.range(rtxn, &(level_0_start..))?.take(first_level_size) {
|
||||||
|
let (_key, (_original_value, docids)) = result?;
|
||||||
|
documents_ids |= docids;
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok((vec![], documents_ids))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn recursive_compute_levels<'t, KeyCodec, ValueCodec, Bound>(
|
||||||
|
rtxn: &'t heed::RoTxn,
|
||||||
|
db: heed::Database<KeyCodec, ValueCodec>,
|
||||||
compression_type: CompressionType,
|
compression_type: CompressionType,
|
||||||
compression_level: Option<u32>,
|
compression_level: Option<u32>,
|
||||||
level: u8,
|
level: u8,
|
||||||
level_0_range: RangeInclusive<(FieldId, u8, f64, f64)>,
|
level_0_start: <KeyCodec as BytesDecode<'t>>::DItem,
|
||||||
|
level_0_range: &'t RangeFrom<<KeyCodec as BytesDecode<'t>>::DItem>,
|
||||||
|
level_0_size: usize,
|
||||||
level_group_size: NonZeroUsize,
|
level_group_size: NonZeroUsize,
|
||||||
computed_group_bitmap: &mut dyn FnMut(&[RoaringBitmap], f64, f64) -> Result<()>,
|
computed_group_bitmap: &mut dyn FnMut(&[RoaringBitmap], Bound, Bound) -> Result<()>,
|
||||||
) -> Result<Vec<Reader<File>>> {
|
bound_from_db_key: &dyn for<'a> Fn(usize, &'a <KeyCodec as BytesDecode<'t>>::DItem) -> Bound,
|
||||||
let (field_id, level_0, first_left, first_right) = level_0_range.start().clone();
|
bitmap_from_db_value: &dyn Fn(<ValueCodec as BytesDecode<'t>>::DItem) -> RoaringBitmap,
|
||||||
assert_eq!(level_0, 0);
|
write_entry: &dyn Fn(&mut Writer<File>, NonZeroU8, Bound, Bound, RoaringBitmap) -> Result<()>,
|
||||||
assert_eq!(first_left, first_right);
|
) -> Result<Vec<Reader<File>>>
|
||||||
|
where
|
||||||
|
KeyCodec: for<'a> BytesEncode<'a>
|
||||||
|
+ for<'a> BytesDecode<'a, DItem = <KeyCodec as BytesEncode<'a>>::EItem>,
|
||||||
|
for<'a> <KeyCodec as BytesEncode<'a>>::EItem: Sized,
|
||||||
|
ValueCodec: for<'a> BytesEncode<'a>
|
||||||
|
+ for<'a> BytesDecode<'a, DItem = <ValueCodec as BytesEncode<'a>>::EItem>,
|
||||||
|
for<'a> <ValueCodec as BytesEncode<'a>>::EItem: Sized,
|
||||||
|
Bound: Copy,
|
||||||
|
{
|
||||||
if level == 0 {
|
if level == 0 {
|
||||||
|
// base case for the recursion
|
||||||
|
|
||||||
let mut bitmaps = vec![];
|
let mut bitmaps = vec![];
|
||||||
|
|
||||||
let mut first_f64_value = first_left;
|
let mut start_bound = bound_from_db_key(0, &level_0_start);
|
||||||
let mut last_f64_value = first_left;
|
let mut end_bound = bound_from_db_key(0, &level_0_start);
|
||||||
|
|
||||||
let mut first_iteration_for_new_group = true;
|
let mut first_iteration_for_new_group = true;
|
||||||
for db_result_item in db.range(rtxn, &level_0_range)? {
|
for (i, db_result_item) in db.range(rtxn, level_0_range)?.take(level_0_size).enumerate() {
|
||||||
let ((_field_id, _level, left, _right), docids) = db_result_item?;
|
let (key, value) = db_result_item?;
|
||||||
// println!("level0: {left}");
|
|
||||||
assert_eq!(_level, 0);
|
let bound = bound_from_db_key(i, &key);
|
||||||
assert_eq!(left, _right);
|
let docids = bitmap_from_db_value(value);
|
||||||
|
|
||||||
if first_iteration_for_new_group {
|
if first_iteration_for_new_group {
|
||||||
first_f64_value = left;
|
start_bound = bound;
|
||||||
first_iteration_for_new_group = false;
|
first_iteration_for_new_group = false;
|
||||||
}
|
}
|
||||||
last_f64_value = left;
|
end_bound = bound;
|
||||||
bitmaps.push(docids);
|
bitmaps.push(docids);
|
||||||
|
|
||||||
if bitmaps.len() == level_group_size.get() {
|
if bitmaps.len() == level_group_size.get() {
|
||||||
// println!("callback first level with {bitmaps:?} {last_f64_value:?}");
|
computed_group_bitmap(&bitmaps, start_bound, end_bound)?;
|
||||||
computed_group_bitmap(&bitmaps, first_f64_value, last_f64_value)?;
|
|
||||||
first_iteration_for_new_group = true;
|
first_iteration_for_new_group = true;
|
||||||
bitmaps.clear();
|
bitmaps.clear();
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if !bitmaps.is_empty() {
|
if !bitmaps.is_empty() {
|
||||||
// println!("end callback first level with {bitmaps:?} {last_f64_value:?}");
|
computed_group_bitmap(&bitmaps, start_bound, end_bound)?;
|
||||||
computed_group_bitmap(&bitmaps, first_f64_value, last_f64_value)?;
|
|
||||||
bitmaps.clear();
|
bitmaps.clear();
|
||||||
}
|
}
|
||||||
|
// level 0 is already stored in the DB
|
||||||
// level 0 isn't actually stored in this DB, since it contains exactly the same information as that other DB
|
|
||||||
return Ok(vec![]);
|
return Ok(vec![]);
|
||||||
} else {
|
} else {
|
||||||
let mut cur_writer =
|
let mut cur_writer =
|
||||||
@ -315,7 +327,9 @@ fn recursive_compute_levels<'t>(
|
|||||||
compression_type,
|
compression_type,
|
||||||
compression_level,
|
compression_level,
|
||||||
level - 1,
|
level - 1,
|
||||||
|
level_0_start,
|
||||||
level_0_range,
|
level_0_range,
|
||||||
|
level_0_size,
|
||||||
level_group_size,
|
level_group_size,
|
||||||
&mut |sub_bitmaps: &[RoaringBitmap], start_range, end_range| {
|
&mut |sub_bitmaps: &[RoaringBitmap], start_range, end_range| {
|
||||||
let mut combined_bitmap = RoaringBitmap::default();
|
let mut combined_bitmap = RoaringBitmap::default();
|
||||||
@ -326,36 +340,33 @@ fn recursive_compute_levels<'t>(
|
|||||||
|
|
||||||
bitmaps.push(combined_bitmap);
|
bitmaps.push(combined_bitmap);
|
||||||
if bitmaps.len() == level_group_size.get() {
|
if bitmaps.len() == level_group_size.get() {
|
||||||
let start_range = range_for_bitmaps.first().unwrap().0;
|
let start_bound = range_for_bitmaps.first().unwrap().0;
|
||||||
let end_range = range_for_bitmaps.last().unwrap().1;
|
let end_bound = range_for_bitmaps.last().unwrap().1;
|
||||||
// println!("callback level {} with {bitmaps:?} {last_f64_value:?}", level + 1);
|
computed_group_bitmap(&bitmaps, start_bound, end_bound)?;
|
||||||
computed_group_bitmap(&bitmaps, start_range, end_range)?;
|
for (bitmap, (start_bound, end_bound)) in
|
||||||
for (bitmap, (start_range, end_range)) in
|
|
||||||
bitmaps.drain(..).zip(range_for_bitmaps.drain(..))
|
bitmaps.drain(..).zip(range_for_bitmaps.drain(..))
|
||||||
{
|
{
|
||||||
// println!("write {field_id} {level} {start_range} {end_range} {bitmap:?}");
|
write_entry(
|
||||||
write_number_entry(
|
|
||||||
&mut cur_writer,
|
&mut cur_writer,
|
||||||
field_id,
|
NonZeroU8::new(level).unwrap(),
|
||||||
level,
|
start_bound,
|
||||||
start_range,
|
end_bound,
|
||||||
end_range,
|
bitmap,
|
||||||
&bitmap,
|
|
||||||
)?;
|
)?;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// println!("end callback level {level}");
|
|
||||||
Ok(())
|
Ok(())
|
||||||
},
|
},
|
||||||
|
bound_from_db_key,
|
||||||
|
bitmap_from_db_value,
|
||||||
|
write_entry,
|
||||||
)?;
|
)?;
|
||||||
if !bitmaps.is_empty() {
|
if !bitmaps.is_empty() {
|
||||||
let start_range = range_for_bitmaps.first().unwrap().0;
|
let start_range = range_for_bitmaps.first().unwrap().0;
|
||||||
let end_range = range_for_bitmaps.last().unwrap().1;
|
let end_range = range_for_bitmaps.last().unwrap().1;
|
||||||
// println!("end callback level {} with {bitmaps:?} {last_f64_value:?}", level + 1);
|
|
||||||
computed_group_bitmap(&bitmaps, start_range, end_range)?;
|
computed_group_bitmap(&bitmaps, start_range, end_range)?;
|
||||||
for (bitmap, (left, right)) in bitmaps.drain(..).zip(range_for_bitmaps.drain(..)) {
|
for (bitmap, (left, right)) in bitmaps.drain(..).zip(range_for_bitmaps.drain(..)) {
|
||||||
// println!("end write: {field_id} {level} {left} {right} {bitmap:?}");
|
write_entry(&mut cur_writer, NonZeroU8::new(level).unwrap(), left, right, bitmap)?;
|
||||||
write_number_entry(&mut cur_writer, field_id, level, left, right, &bitmap)?;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -364,113 +375,15 @@ fn recursive_compute_levels<'t>(
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn compute_facet_number_levels<'t>(
|
fn clear_field_number_levels<'t>(
|
||||||
rtxn: &'t heed::RoTxn,
|
wtxn: &'t mut heed::RwTxn,
|
||||||
db: heed::Database<FacetLevelValueF64Codec, CboRoaringBitmapCodec>,
|
db: heed::Database<FacetLevelValueF64Codec, CboRoaringBitmapCodec>,
|
||||||
compression_type: CompressionType,
|
|
||||||
compression_level: Option<u32>,
|
|
||||||
level_group_size: NonZeroUsize,
|
|
||||||
min_level_size: NonZeroUsize,
|
|
||||||
field_id: FieldId,
|
field_id: FieldId,
|
||||||
) -> Result<Reader<File>> {
|
) -> heed::Result<()> {
|
||||||
let first_level_size = db
|
let left = (field_id, 1, f64::MIN, f64::MIN);
|
||||||
.remap_key_type::<ByteSlice>()
|
let right = (field_id, u8::MAX, f64::MAX, f64::MAX);
|
||||||
.prefix_iter(rtxn, &field_id.to_be_bytes())?
|
let range = left..=right;
|
||||||
.remap_types::<DecodeIgnore, DecodeIgnore>()
|
db.delete_range(wtxn, &range).map(drop)
|
||||||
.fold(Ok(0usize), |count, result| result.and(count).map(|c| c + 1))?;
|
|
||||||
|
|
||||||
// It is forbidden to keep a cursor and write in a database at the same time with LMDB
|
|
||||||
// therefore we write the facet levels entries into a grenad file before transfering them.
|
|
||||||
let mut writer = create_writer(compression_type, compression_level, tempfile::tempfile()?);
|
|
||||||
|
|
||||||
let level_0_range = {
|
|
||||||
let left = (field_id, 0, f64::MIN, f64::MIN);
|
|
||||||
let right = (field_id, 0, f64::MAX, f64::MAX);
|
|
||||||
left..=right
|
|
||||||
};
|
|
||||||
|
|
||||||
// Groups sizes are always a power of the original level_group_size and therefore a group
|
|
||||||
// always maps groups of the previous level and never splits previous levels groups in half.
|
|
||||||
let group_size_iter = (1u8..)
|
|
||||||
.map(|l| (l, level_group_size.get().pow(l as u32)))
|
|
||||||
.take_while(|(_, s)| first_level_size / *s >= min_level_size.get());
|
|
||||||
|
|
||||||
for (level, group_size) in group_size_iter {
|
|
||||||
// dbg!(level, group_size);
|
|
||||||
let mut left = 0.0;
|
|
||||||
let mut right = 0.0;
|
|
||||||
let mut group_docids = RoaringBitmap::new();
|
|
||||||
|
|
||||||
for (i, result) in db.range(rtxn, &level_0_range)?.enumerate() {
|
|
||||||
let ((_field_id, _level, value, _right), docids) = result?;
|
|
||||||
|
|
||||||
if i == 0 {
|
|
||||||
left = value;
|
|
||||||
} else if i % group_size == 0 {
|
|
||||||
// we found the first bound of the next group, we must store the left
|
|
||||||
// and right bounds associated with the docids.
|
|
||||||
write_number_entry(&mut writer, field_id, level, left, right, &group_docids)?;
|
|
||||||
|
|
||||||
// We save the left bound for the new group and also reset the docids.
|
|
||||||
group_docids = RoaringBitmap::new();
|
|
||||||
left = value;
|
|
||||||
}
|
|
||||||
|
|
||||||
// The right bound is always the bound we run through.
|
|
||||||
group_docids |= docids;
|
|
||||||
right = value;
|
|
||||||
}
|
|
||||||
|
|
||||||
if !group_docids.is_empty() {
|
|
||||||
write_number_entry(&mut writer, field_id, level, left, right, &group_docids)?;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
writer_into_reader(writer)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn write_number_entry(
|
|
||||||
writer: &mut Writer<File>,
|
|
||||||
field_id: FieldId,
|
|
||||||
level: u8,
|
|
||||||
left: f64,
|
|
||||||
right: f64,
|
|
||||||
ids: &RoaringBitmap,
|
|
||||||
) -> Result<()> {
|
|
||||||
let key = (field_id, level, left, right);
|
|
||||||
let key = FacetLevelValueF64Codec::bytes_encode(&key).ok_or(Error::Encoding)?;
|
|
||||||
let data = CboRoaringBitmapCodec::bytes_encode(&ids).ok_or(Error::Encoding)?;
|
|
||||||
// println!(" w{field_id}-{level}-{left}-{right}");
|
|
||||||
writer.insert(&key, &data)?;
|
|
||||||
Ok(())
|
|
||||||
}
|
|
||||||
|
|
||||||
fn compute_faceted_strings_documents_ids(
|
|
||||||
rtxn: &heed::RoTxn,
|
|
||||||
db: heed::Database<ByteSlice, FacetStringLevelZeroValueCodec>,
|
|
||||||
field_id: FieldId,
|
|
||||||
) -> Result<RoaringBitmap> {
|
|
||||||
let mut documents_ids = RoaringBitmap::new();
|
|
||||||
for result in db.prefix_iter(rtxn, &field_id.to_be_bytes())? {
|
|
||||||
let (_key, (_original_value, docids)) = result?;
|
|
||||||
documents_ids |= docids;
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(documents_ids)
|
|
||||||
}
|
|
||||||
|
|
||||||
fn compute_faceted_numbers_documents_ids(
|
|
||||||
rtxn: &heed::RoTxn,
|
|
||||||
db: heed::Database<ByteSlice, CboRoaringBitmapCodec>,
|
|
||||||
field_id: FieldId,
|
|
||||||
) -> Result<RoaringBitmap> {
|
|
||||||
let mut documents_ids = RoaringBitmap::new();
|
|
||||||
for result in db.prefix_iter(rtxn, &field_id.to_be_bytes())? {
|
|
||||||
let (_key, docids) = result?;
|
|
||||||
documents_ids |= docids;
|
|
||||||
}
|
|
||||||
|
|
||||||
Ok(documents_ids)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn clear_field_string_levels<'t>(
|
fn clear_field_string_levels<'t>(
|
||||||
@ -484,68 +397,20 @@ fn clear_field_string_levels<'t>(
|
|||||||
db.remap_key_type::<FacetLevelValueU32Codec>().delete_range(wtxn, &range).map(drop)
|
db.remap_key_type::<FacetLevelValueU32Codec>().delete_range(wtxn, &range).map(drop)
|
||||||
}
|
}
|
||||||
|
|
||||||
fn compute_facet_string_levels<'t>(
|
fn write_number_entry(
|
||||||
rtxn: &'t heed::RoTxn,
|
writer: &mut Writer<File>,
|
||||||
db: heed::Database<FacetStringLevelZeroCodec, FacetStringLevelZeroValueCodec>,
|
|
||||||
compression_type: CompressionType,
|
|
||||||
compression_level: Option<u32>,
|
|
||||||
level_group_size: NonZeroUsize,
|
|
||||||
min_level_size: NonZeroUsize,
|
|
||||||
field_id: FieldId,
|
field_id: FieldId,
|
||||||
) -> Result<Reader<File>> {
|
level: u8,
|
||||||
let first_level_size = db
|
left: f64,
|
||||||
.remap_key_type::<ByteSlice>()
|
right: f64,
|
||||||
.prefix_iter(rtxn, &field_id.to_be_bytes())?
|
ids: &RoaringBitmap,
|
||||||
.remap_types::<DecodeIgnore, DecodeIgnore>()
|
) -> Result<()> {
|
||||||
.fold(Ok(0usize), |count, result| result.and(count).map(|c| c + 1))?;
|
let key = (field_id, level, left, right);
|
||||||
|
let key = FacetLevelValueF64Codec::bytes_encode(&key).ok_or(Error::Encoding)?;
|
||||||
// It is forbidden to keep a cursor and write in a database at the same time with LMDB
|
let data = CboRoaringBitmapCodec::bytes_encode(&ids).ok_or(Error::Encoding)?;
|
||||||
// therefore we write the facet levels entries into a grenad file before transfering them.
|
writer.insert(&key, &data)?;
|
||||||
let mut writer = create_writer(compression_type, compression_level, tempfile::tempfile()?);
|
Ok(())
|
||||||
|
|
||||||
// Groups sizes are always a power of the original level_group_size and therefore a group
|
|
||||||
// always maps groups of the previous level and never splits previous levels groups in half.
|
|
||||||
let group_size_iter = (1u8..)
|
|
||||||
.map(|l| (l, level_group_size.get().pow(l as u32)))
|
|
||||||
.take_while(|(_, s)| first_level_size / *s >= min_level_size.get());
|
|
||||||
|
|
||||||
for (level, group_size) in group_size_iter {
|
|
||||||
let level = NonZeroU8::new(level).unwrap();
|
|
||||||
let mut left = (0, "");
|
|
||||||
let mut right = (0, "");
|
|
||||||
let mut group_docids = RoaringBitmap::new();
|
|
||||||
|
|
||||||
// Because we know the size of the level 0 we can use a range iterator that starts
|
|
||||||
// at the first value of the level and goes to the last by simply counting.
|
|
||||||
for (i, result) in db.range(rtxn, &((field_id, "")..))?.take(first_level_size).enumerate() {
|
|
||||||
let ((_field_id, value), (_original_value, docids)) = result?;
|
|
||||||
|
|
||||||
if i == 0 {
|
|
||||||
left = (i as u32, value);
|
|
||||||
} else if i % group_size == 0 {
|
|
||||||
// we found the first bound of the next group, we must store the left
|
|
||||||
// and right bounds associated with the docids. We also reset the docids.
|
|
||||||
let docids = mem::take(&mut group_docids);
|
|
||||||
write_string_entry(&mut writer, field_id, level, left, right, docids)?;
|
|
||||||
|
|
||||||
// We save the left bound for the new group.
|
|
||||||
left = (i as u32, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
// The right bound is always the bound we run through.
|
|
||||||
group_docids |= docids;
|
|
||||||
right = (i as u32, value);
|
|
||||||
}
|
|
||||||
|
|
||||||
if !group_docids.is_empty() {
|
|
||||||
let docids = mem::take(&mut group_docids);
|
|
||||||
write_string_entry(&mut writer, field_id, level, left, right, docids)?;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
writer_into_reader(writer)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fn write_string_entry(
|
fn write_string_entry(
|
||||||
writer: &mut Writer<File>,
|
writer: &mut Writer<File>,
|
||||||
field_id: FieldId,
|
field_id: FieldId,
|
||||||
|
Loading…
Reference in New Issue
Block a user