mirror of
https://github.com/meilisearch/MeiliSearch
synced 2025-07-04 20:37:15 +02:00
Introduce the enable-chunk-fusing flag
This commit is contained in:
parent
f980422c57
commit
9021b2dba6
3 changed files with 37 additions and 18 deletions
|
@ -110,8 +110,17 @@ struct IndexerOpt {
|
|||
#[structopt(long, requires = "chunk-compression-type")]
|
||||
chunk_compression_level: Option<u32>,
|
||||
|
||||
/// The number of bytes to remove from the begining of the chunks while reading/sorting
|
||||
/// or merging them.
|
||||
///
|
||||
/// File fusing must only be enable on file systems that support the `FALLOC_FL_COLLAPSE_RANGE`,
|
||||
/// (i.e. ext4 and XFS). File fusing will only work if the `enable-chunk-fusing` is set.
|
||||
#[structopt(long, default_value = "4294967296")] // 4 GB
|
||||
file_fusing_shrink_size: u64,
|
||||
chunk_fusing_shrink_size: u64,
|
||||
|
||||
/// Enable the chunk fusing or not, this reduces the amount of disk used by a factor of 2.
|
||||
#[structopt(long)]
|
||||
enable_chunk_fusing: bool,
|
||||
}
|
||||
|
||||
fn format_count(n: usize) -> String {
|
||||
|
@ -131,10 +140,14 @@ fn create_writer(typ: CompressionType, level: Option<u32>, file: File) -> io::Re
|
|||
builder.build(file)
|
||||
}
|
||||
|
||||
fn writer_into_reader(writer: Writer<File>, shrink_size: u64) -> anyhow::Result<Reader<FileFuse>> {
|
||||
fn writer_into_reader(writer: Writer<File>, shrink_size: Option<u64>) -> anyhow::Result<Reader<FileFuse>> {
|
||||
let mut file = writer.into_inner()?;
|
||||
file.seek(SeekFrom::Start(0))?;
|
||||
let file = FileFuse::with_shrink_size(file, shrink_size);
|
||||
let file = if let Some(shrink_size) = shrink_size {
|
||||
FileFuse::builder().shrink_size(shrink_size).build(file)
|
||||
} else {
|
||||
FileFuse::new(file)
|
||||
};
|
||||
Reader::new(file).map_err(Into::into)
|
||||
}
|
||||
|
||||
|
@ -142,13 +155,15 @@ fn create_sorter(
|
|||
merge: MergeFn,
|
||||
chunk_compression_type: CompressionType,
|
||||
chunk_compression_level: Option<u32>,
|
||||
file_fusing_shrink_size: u64,
|
||||
chunk_fusing_shrink_size: Option<u64>,
|
||||
max_nb_chunks: Option<usize>,
|
||||
max_memory: Option<usize>,
|
||||
) -> Sorter<MergeFn>
|
||||
{
|
||||
let mut builder = Sorter::builder(merge);
|
||||
builder.file_fusing_shrink_size(file_fusing_shrink_size);
|
||||
if let Some(shrink_size) = chunk_fusing_shrink_size {
|
||||
builder.file_fusing_shrink_size(shrink_size);
|
||||
}
|
||||
builder.chunk_compression_type(chunk_compression_type);
|
||||
if let Some(level) = chunk_compression_level {
|
||||
builder.chunk_compression_level(level);
|
||||
|
@ -215,7 +230,7 @@ struct Store {
|
|||
// MTBL parameters
|
||||
chunk_compression_type: CompressionType,
|
||||
chunk_compression_level: Option<u32>,
|
||||
file_fusing_shrink_size: u64,
|
||||
chunk_fusing_shrink_size: Option<u64>,
|
||||
// MTBL sorters
|
||||
main_sorter: Sorter<MergeFn>,
|
||||
word_docids_sorter: Sorter<MergeFn>,
|
||||
|
@ -232,7 +247,7 @@ impl Store {
|
|||
max_memory: Option<usize>,
|
||||
chunk_compression_type: CompressionType,
|
||||
chunk_compression_level: Option<u32>,
|
||||
file_fusing_shrink_size: u64,
|
||||
chunk_fusing_shrink_size: Option<u64>,
|
||||
) -> anyhow::Result<Store>
|
||||
{
|
||||
// We divide the max memory by the number of sorter the Store have.
|
||||
|
@ -242,7 +257,7 @@ impl Store {
|
|||
main_merge,
|
||||
chunk_compression_type,
|
||||
chunk_compression_level,
|
||||
file_fusing_shrink_size,
|
||||
chunk_fusing_shrink_size,
|
||||
max_nb_chunks,
|
||||
max_memory,
|
||||
);
|
||||
|
@ -250,7 +265,7 @@ impl Store {
|
|||
word_docids_merge,
|
||||
chunk_compression_type,
|
||||
chunk_compression_level,
|
||||
file_fusing_shrink_size,
|
||||
chunk_fusing_shrink_size,
|
||||
max_nb_chunks,
|
||||
max_memory,
|
||||
);
|
||||
|
@ -258,7 +273,7 @@ impl Store {
|
|||
words_pairs_proximities_docids_merge,
|
||||
chunk_compression_type,
|
||||
chunk_compression_level,
|
||||
file_fusing_shrink_size,
|
||||
chunk_fusing_shrink_size,
|
||||
max_nb_chunks,
|
||||
max_memory,
|
||||
);
|
||||
|
@ -278,7 +293,7 @@ impl Store {
|
|||
documents_ids: RoaringBitmap::new(),
|
||||
chunk_compression_type,
|
||||
chunk_compression_level,
|
||||
file_fusing_shrink_size,
|
||||
chunk_fusing_shrink_size,
|
||||
|
||||
main_sorter,
|
||||
word_docids_sorter,
|
||||
|
@ -521,7 +536,7 @@ impl Store {
|
|||
fn finish(mut self) -> anyhow::Result<Readers> {
|
||||
let comp_type = self.chunk_compression_type;
|
||||
let comp_level = self.chunk_compression_level;
|
||||
let shrink_size = self.file_fusing_shrink_size;
|
||||
let shrink_size = self.chunk_fusing_shrink_size;
|
||||
|
||||
Self::write_word_docids(&mut self.word_docids_sorter, self.word_docids)?;
|
||||
Self::write_documents_ids(&mut self.main_sorter, self.documents_ids)?;
|
||||
|
@ -754,9 +769,14 @@ fn main() -> anyhow::Result<()> {
|
|||
let max_memory_by_job = opt.indexer.max_memory / num_threads;
|
||||
let chunk_compression_type = opt.indexer.chunk_compression_type;
|
||||
let chunk_compression_level = opt.indexer.chunk_compression_level;
|
||||
let file_fusing_shrink_size = opt.indexer.file_fusing_shrink_size;
|
||||
let log_every_n = opt.indexer.log_every_n;
|
||||
|
||||
let chunk_fusing_shrink_size = if opt.indexer.enable_chunk_fusing {
|
||||
Some(opt.indexer.chunk_fusing_shrink_size)
|
||||
} else {
|
||||
None
|
||||
};
|
||||
|
||||
let readers = csv_readers(opt.csv_file, num_threads)?
|
||||
.into_par_iter()
|
||||
.enumerate()
|
||||
|
@ -767,7 +787,7 @@ fn main() -> anyhow::Result<()> {
|
|||
Some(max_memory_by_job),
|
||||
chunk_compression_type,
|
||||
chunk_compression_level,
|
||||
file_fusing_shrink_size,
|
||||
chunk_fusing_shrink_size,
|
||||
)?;
|
||||
store.index_csv(rdr, i, num_threads, log_every_n)
|
||||
})
|
||||
|
@ -794,7 +814,7 @@ fn main() -> anyhow::Result<()> {
|
|||
})?;
|
||||
let merger = merge_readers(readers, merge);
|
||||
merger.write_into(&mut writer)?;
|
||||
writer_into_reader(writer, file_fusing_shrink_size)
|
||||
writer_into_reader(writer, chunk_fusing_shrink_size)
|
||||
};
|
||||
|
||||
// The enum and the channel which is used to transfert
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue