mod errors; use meili_snap::insta::assert_json_snapshot; use meili_snap::snapshot; use time::format_description::well_known::Rfc3339; use time::OffsetDateTime; use crate::common::Server; use crate::json; #[actix_rt::test] async fn error_get_unexisting_batch_status() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; let (response, code) = index.get_batch(1).await; let expected_response = json!({ "message": "batch `1` not found.", "code": "batch_not_found", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#batch_not_found" }); assert_eq!(response, expected_response); assert_eq!(code, 404); } #[actix_rt::test] async fn get_batch_status() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index .add_documents( json!([{ "id": 1, "content": "foobar", }]), None, ) .await; index.wait_task(0).await; let (_response, code) = index.get_batch(1).await; assert_eq!(code, 200); // TODO check response format, as per #48 } #[actix_rt::test] async fn list_batches() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.list_batches().await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_with_star_filters() { let server = Server::new().await; let index = server.index("test"); let (batch, _code) = index.create(None).await; index.wait_task(batch.uid()).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.service.get("/batches?indexUids=test").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?indexUids=*").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?indexUids=*,pasteque").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?types=*").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?types=*,documentAdditionOrUpdate&statuses=*").await; assert_eq!(code, 200, "{:?}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index .service .get("/batches?types=*,documentAdditionOrUpdate&statuses=*,failed&indexUids=test") .await; assert_eq!(code, 200, "{:?}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index .service .get("/batches?types=*,documentAdditionOrUpdate&statuses=*,failed&indexUids=test,*") .await; assert_eq!(code, 200, "{:?}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_status_filtered() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&[], &["succeeded"], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 1); // We can't be sure that the update isn't already processed so we can't test this // let (response, code) = index.filtered_batches(&[], &["processing"]).await; // assert_eq!(code, 200, "{}", response); // assert_eq!(response["results"].as_array().unwrap().len(), 1); index.wait_task(1).await; let (response, code) = index.filtered_batches(&[], &["succeeded"], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_type_filtered() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&["indexCreation"], &[], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 1); let (response, code) = index.filtered_batches(&["indexCreation", "documentAdditionOrUpdate"], &[], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_invalid_canceled_by_filter() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&[], &[], &["0"]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 0); } #[actix_rt::test] async fn list_batches_status_and_type_filtered() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&["indexCreation"], &["failed"], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 0); let (response, code) = index .filtered_batches( &["indexCreation", "documentAdditionOrUpdate"], &["succeeded", "processing", "enqueued"], &[], ) .await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn get_batch_filter_error() { let server = Server::new().await; let (response, code) = server.batches_filter("lol=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r###" { "message": "Unknown parameter `lol`: expected one of `limit`, `from`, `batchUids`, `uids`, `canceledBy`, `types`, `statuses`, `indexUids`, `afterEnqueuedAt`, `beforeEnqueuedAt`, `afterStartedAt`, `beforeStartedAt`, `afterFinishedAt`, `beforeFinishedAt`", "code": "bad_request", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#bad_request" } "###); let (response, code) = server.batches_filter("uids=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r#" { "message": "Invalid value in parameter `uids`: could not parse `pied` as a positive integer", "code": "invalid_task_uids", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_task_uids" } "#); let (response, code) = server.batches_filter("from=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r#" { "message": "Invalid value in parameter `from`: could not parse `pied` as a positive integer", "code": "invalid_task_from", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_task_from" } "#); let (response, code) = server.batches_filter("beforeStartedAt=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r#" { "message": "Invalid value in parameter `beforeStartedAt`: `pied` is an invalid date-time. It should follow the YYYY-MM-DD or RFC 3339 date-time format.", "code": "invalid_task_before_started_at", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_task_before_started_at" } "#); } macro_rules! assert_valid_summarized_batch { ($response:expr, $batch_type:literal, $index:literal) => {{ assert_eq!($response.as_object().unwrap().len(), 5); assert!($response["batchUid"].as_u64().is_some()); assert_eq!($response["indexUid"], $index); assert_eq!($response["status"], "enqueued"); assert_eq!($response["type"], $batch_type); let date = $response["enqueuedAt"].as_str().expect("missing date"); OffsetDateTime::parse(date, &Rfc3339).unwrap(); }}; } #[actix_web::test] async fn test_summarized_batch_view() { let server = Server::new().await; let index = server.index("test"); let (response, _) = index.create(None).await; assert_valid_summarized_batch!(response, "indexCreation", "test"); let (response, _) = index.update(None).await; assert_valid_summarized_batch!(response, "indexUpdate", "test"); let (response, _) = index.update_settings(json!({})).await; assert_valid_summarized_batch!(response, "settingsUpdate", "test"); let (response, _) = index.update_documents(json!([{"id": 1}]), None).await; assert_valid_summarized_batch!(response, "documentAdditionOrUpdate", "test"); let (response, _) = index.add_documents(json!([{"id": 1}]), None).await; assert_valid_summarized_batch!(response, "documentAdditionOrUpdate", "test"); let (response, _) = index.delete_document(1).await; assert_valid_summarized_batch!(response, "documentDeletion", "test"); let (response, _) = index.clear_all_documents().await; assert_valid_summarized_batch!(response, "documentDeletion", "test"); let (response, _) = index.delete().await; assert_valid_summarized_batch!(response, "indexDeletion", "test"); } #[actix_web::test] async fn test_summarized_document_addition_or_update() { let server = Server::new().await; let index = server.index("test"); index.add_documents(json!({ "id": 42, "content": "doggos & fluff" }), None).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "receivedDocuments": 1, "indexedDocuments": 1 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentAdditionOrUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.add_documents(json!({ "id": 42, "content": "doggos & fluff" }), Some("id")).await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "receivedDocuments": 1, "indexedDocuments": 1 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentAdditionOrUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_delete_documents_by_batch() { let server = Server::new().await; let index = server.index("test"); index.delete_batch(vec![1, 2, 3]).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(None).await; index.delete_batch(vec![42]).await; index.wait_task(2).await; let (batch, _) = index.get_batch(2).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 2, "details": { "providedIds": 1, "deletedDocuments": 0 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_delete_documents_by_filter() { let server = Server::new().await; let index = server.index("test"); index.delete_document_by_filter(json!({ "filter": "doggo = bernese" })).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(None).await; index.delete_document_by_filter(json!({ "filter": "doggo = bernese" })).await; index.wait_task(2).await; let (batch, _) = index.get_batch(2).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 2, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.update_settings(json!({ "filterableAttributes": ["doggo"] })).await; index.delete_document_by_filter(json!({ "filter": "doggo = bernese" })).await; index.wait_task(4).await; let (batch, _) = index.get_batch(4).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 4, "details": { "providedIds": 0, "deletedDocuments": 0, "originalFilter": "\"doggo = bernese\"" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_delete_document_by_id() { let server = Server::new().await; let index = server.index("test"); index.delete_document(1).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(None).await; index.delete_document(42).await; index.wait_task(2).await; let (batch, _) = index.get_batch(2).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 2, "details": { "providedIds": 1, "deletedDocuments": 0 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_settings_update() { let server = Server::new().await; let index = server.index("test"); // here we should find my payload even in the failed batch. let (response, code) = index.update_settings(json!({ "rankingRules": ["custom"] })).await; meili_snap::snapshot!(code, @"400 Bad Request"); meili_snap::snapshot!(meili_snap::json_string!(response), @r###" { "message": "Invalid value at `.rankingRules[0]`: `custom` ranking rule is invalid. Valid ranking rules are words, typo, sort, proximity, attribute, exactness and custom ranking rules.", "code": "invalid_settings_ranking_rules", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_settings_ranking_rules" } "###); index.update_settings(json!({ "displayedAttributes": ["doggos", "name"], "filterableAttributes": ["age", "nb_paw_pads"], "sortableAttributes": ["iq"] })).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "displayedAttributes": [ "doggos", "name" ], "filterableAttributes": [ "age", "nb_paw_pads" ], "sortableAttributes": [ "iq" ] }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "settingsUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_index_creation() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexCreation": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(Some("doggos")).await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexCreation": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_index_deletion() { let server = Server::new().await; let index = server.index("test"); let (ret, _code) = index.delete().await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "failed", "type": "indexDeletion", "canceledBy": null, "details": { "deletedDocuments": 0 }, "error": { "message": "Index `test` not found.", "code": "index_not_found", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#index_not_found" }, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); // is the details correctly set when documents are actually deleted. // /!\ We need to wait for the document addition to be processed otherwise, if the test runs too slow, // both batches may get autobatched and the deleted documents count will be wrong. let (ret, _code) = index.add_documents(json!({ "id": 42, "content": "doggos & fluff" }), Some("id")).await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "succeeded", "type": "documentAdditionOrUpdate", "canceledBy": null, "details": { "receivedDocuments": 1, "indexedDocuments": 1 }, "error": null, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); let (ret, _code) = index.delete().await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "succeeded", "type": "indexDeletion", "canceledBy": null, "details": { "deletedDocuments": 1 }, "error": null, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); // What happens when you delete an index that doesn't exists. let (ret, _code) = index.delete().await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "failed", "type": "indexDeletion", "canceledBy": null, "details": { "deletedDocuments": 0 }, "error": { "message": "Index `test` not found.", "code": "index_not_found", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#index_not_found" }, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); } #[actix_web::test] async fn test_summarized_index_update() { let server = Server::new().await; let index = server.index("test"); // If the index doesn't exist yet, we should get errors with or without the primary key. index.update(None).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.update(Some("bones")).await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); // And run the same two tests once the index do exists. index.create(None).await; index.update(None).await; index.wait_task(3).await; let (batch, _) = index.get_batch(3).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 3, "details": {}, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.update(Some("bones")).await; index.wait_task(4).await; let (batch, _) = index.get_batch(4).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 4, "details": { "primaryKey": "bones" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_index_swap() { let server = Server::new().await; server .index_swap(json!([ { "indexes": ["doggos", "cattos"] } ])) .await; server.wait_task(0).await; let (batch, _) = server.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexSwap": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); server.index("doggos").create(None).await; server.index("cattos").create(None).await; server .index_swap(json!([ { "indexes": ["doggos", "cattos"] } ])) .await; server.wait_task(3).await; let (batch, _) = server.get_batch(3).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 3, "details": { "swaps": [ { "indexes": [ "doggos", "cattos" ] } ] }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexSwap": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_batch_cancelation() { let server = Server::new().await; let index = server.index("doggos"); // to avoid being flaky we're only going to cancel an already finished batch :( index.create(None).await; index.wait_task(0).await; server.cancel_tasks("uids=0").await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "matchedTasks": 1, "canceledTasks": 0, "originalFilter": "?uids=0" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "taskCancelation": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_batch_deletion() { let server = Server::new().await; let index = server.index("doggos"); // to avoid being flaky we're only going to delete an already finished batch :( index.create(None).await; index.wait_task(0).await; server.delete_tasks("uids=0").await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "matchedTasks": 1, "deletedTasks": 1, "originalFilter": "?uids=0" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "taskDeletion": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_dump_creation() { let server = Server::new().await; server.create_dump().await; server.wait_task(0).await; let (batch, _) = server.get_batch(0).await; assert_json_snapshot!(batch, { ".details.dumpUid" => "[dumpUid]", ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "dumpUid": "[dumpUid]" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "dumpCreation": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); }