diff --git a/crates/index-scheduler/src/scheduler/process_export.rs b/crates/index-scheduler/src/scheduler/process_export.rs index 19b2bf743..d1f5616b7 100644 --- a/crates/index-scheduler/src/scheduler/process_export.rs +++ b/crates/index-scheduler/src/scheduler/process_export.rs @@ -30,7 +30,7 @@ impl IndexScheduler { payload_size: Option<&Byte>, indexes: &BTreeMap, progress: Progress, - ) -> Result> { + ) -> Result> { #[cfg(test)] self.maybe_fail(crate::test_utils::FailureLocation::ProcessExport)?; @@ -48,7 +48,7 @@ impl IndexScheduler { let mut output = BTreeMap::new(); let agent = ureq::AgentBuilder::new().timeout(Duration::from_secs(5)).build(); let must_stop_processing = self.scheduler.must_stop_processing.clone(); - for (i, (pattern, uid, export_settings)) in indexes.iter().enumerate() { + for (i, (_pattern, uid, export_settings)) in indexes.iter().enumerate() { if must_stop_processing.get() { return Err(Error::AbortedTask); } @@ -63,9 +63,8 @@ impl IndexScheduler { let index = self.index(uid)?; let index_rtxn = index.read_txn()?; - let url = format!("{base_url}/indexes/{uid}"); - // First, check if the index already exists + let url = format!("{base_url}/indexes/{uid}"); let response = retry(&must_stop_processing, || { let mut request = agent.get(&url); if let Some(api_key) = api_key { @@ -158,7 +157,7 @@ impl IndexScheduler { progress.update_progress(progress_step); output.insert( - (*pattern).clone(), + uid.clone(), DetailsExportIndexSettings { settings: (*export_settings).clone(), matched_documents: Some(total_documents as u64), diff --git a/crates/meilisearch-types/src/tasks.rs b/crates/meilisearch-types/src/tasks.rs index 99b04f1e3..423cf539e 100644 --- a/crates/meilisearch-types/src/tasks.rs +++ b/crates/meilisearch-types/src/tasks.rs @@ -289,12 +289,12 @@ impl KindWithContent { }), KindWithContent::DumpCreation { .. } => Some(Details::Dump { dump_uid: None }), KindWithContent::SnapshotCreation => None, - KindWithContent::Export { url, api_key, payload_size, indexes } => { + KindWithContent::Export { url, api_key, payload_size, indexes: _ } => { Some(Details::Export { url: url.clone(), api_key: api_key.clone(), payload_size: *payload_size, - indexes: indexes.iter().map(|(p, s)| (p.clone(), s.clone().into())).collect(), + indexes: BTreeMap::new(), }) } KindWithContent::UpgradeDatabase { from } => Some(Details::UpgradeDatabase { @@ -363,12 +363,12 @@ impl KindWithContent { }), KindWithContent::DumpCreation { .. } => Some(Details::Dump { dump_uid: None }), KindWithContent::SnapshotCreation => None, - KindWithContent::Export { url, api_key, payload_size, indexes } => { + KindWithContent::Export { url, api_key, payload_size, indexes: _ } => { Some(Details::Export { url: url.clone(), api_key: api_key.clone(), payload_size: *payload_size, - indexes: indexes.iter().map(|(p, s)| (p.clone(), s.clone().into())).collect(), + indexes: BTreeMap::new(), }) } KindWithContent::UpgradeDatabase { from } => Some(Details::UpgradeDatabase { @@ -419,12 +419,12 @@ impl From<&KindWithContent> for Option
{ }), KindWithContent::DumpCreation { .. } => Some(Details::Dump { dump_uid: None }), KindWithContent::SnapshotCreation => None, - KindWithContent::Export { url, api_key, payload_size, indexes } => { + KindWithContent::Export { url, api_key, payload_size, indexes: _ } => { Some(Details::Export { url: url.clone(), api_key: api_key.clone(), payload_size: *payload_size, - indexes: indexes.iter().map(|(p, s)| (p.clone(), s.clone().into())).collect(), + indexes: BTreeMap::new(), }) } KindWithContent::UpgradeDatabase { from } => Some(Details::UpgradeDatabase { @@ -696,7 +696,7 @@ pub enum Details { url: String, api_key: Option, payload_size: Option, - indexes: BTreeMap, + indexes: BTreeMap, }, UpgradeDatabase { from: (u32, u32, u32),