mod errors; use meili_snap::insta::assert_json_snapshot; use meili_snap::snapshot; use crate::common::Server; use crate::json; #[actix_rt::test] async fn error_get_unexisting_batch_status() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; let (response, code) = index.get_batch(1).await; let expected_response = json!({ "message": "Batch `1` not found.", "code": "batch_not_found", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#batch_not_found" }); assert_eq!(response, expected_response); assert_eq!(code, 404); } #[actix_rt::test] async fn get_batch_status() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; let (_response, code) = index.get_batch(0).await; assert_eq!(code, 200); } #[actix_rt::test] async fn list_batches() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.list_batches().await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_pagination_and_reverse() { let server = Server::new().await; // First of all we want to create a lot of batches very quickly. The fastest way is to delete a lot of unexisting indexes let mut last_batch = None; for i in 0..10 { let index = server.index(format!("test-{i}")); last_batch = Some(index.create(None).await.0.uid()); } server.wait_task(last_batch.unwrap()).await; let (response, code) = server.batches_filter("limit=3").await; assert_eq!(code, 200); let results = response["results"].as_array().unwrap(); let batch_ids: Vec<_> = results.iter().map(|ret| ret["uid"].as_u64().unwrap()).collect(); snapshot!(format!("{batch_ids:?}"), @"[9, 8, 7]"); let (response, code) = server.batches_filter("limit=3&from=1").await; assert_eq!(code, 200); let results = response["results"].as_array().unwrap(); let batch_ids: Vec<_> = results.iter().map(|ret| ret["uid"].as_u64().unwrap()).collect(); snapshot!(format!("{batch_ids:?}"), @"[1, 0]"); // In reversed order let (response, code) = server.batches_filter("limit=3&reverse=true").await; assert_eq!(code, 200); let results = response["results"].as_array().unwrap(); let batch_ids: Vec<_> = results.iter().map(|ret| ret["uid"].as_u64().unwrap()).collect(); snapshot!(format!("{batch_ids:?}"), @"[0, 1, 2]"); let (response, code) = server.batches_filter("limit=3&from=8&reverse=true").await; assert_eq!(code, 200); let results = response["results"].as_array().unwrap(); let batch_ids: Vec<_> = results.iter().map(|ret| ret["uid"].as_u64().unwrap()).collect(); snapshot!(format!("{batch_ids:?}"), @"[8, 9]"); } #[actix_rt::test] async fn list_batches_with_star_filters() { let server = Server::new().await; let index = server.index("test"); let (batch, _code) = index.create(None).await; index.wait_task(batch.uid()).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.service.get("/batches?indexUids=test").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?indexUids=*").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?indexUids=*,pasteque").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?types=*").await; assert_eq!(code, 200); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index.service.get("/batches?types=*,documentAdditionOrUpdate&statuses=*").await; assert_eq!(code, 200, "{:?}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index .service .get("/batches?types=*,documentAdditionOrUpdate&statuses=*,failed&indexUids=test") .await; assert_eq!(code, 200, "{:?}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); let (response, code) = index .service .get("/batches?types=*,documentAdditionOrUpdate&statuses=*,failed&indexUids=test,*") .await; assert_eq!(code, 200, "{:?}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_status_filtered() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&[], &["succeeded"], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 1); // We can't be sure that the update isn't already processed so we can't test this // let (response, code) = index.filtered_batches(&[], &["processing"]).await; // assert_eq!(code, 200, "{}", response); // assert_eq!(response["results"].as_array().unwrap().len(), 1); index.wait_task(1).await; let (response, code) = index.filtered_batches(&[], &["succeeded"], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_type_filtered() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&["indexCreation"], &[], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 1); let (response, code) = index.filtered_batches(&["indexCreation", "documentAdditionOrUpdate"], &[], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn list_batches_invalid_canceled_by_filter() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&[], &[], &["0"]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 0); } #[actix_rt::test] async fn list_batches_status_and_type_filtered() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; index .add_documents(serde_json::from_str(include_str!("../assets/test_set.json")).unwrap(), None) .await; let (response, code) = index.filtered_batches(&["indexCreation"], &["failed"], &[]).await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 0); let (response, code) = index .filtered_batches( &["indexCreation", "documentAdditionOrUpdate"], &["succeeded", "processing", "enqueued"], &[], ) .await; assert_eq!(code, 200, "{}", response); assert_eq!(response["results"].as_array().unwrap().len(), 2); } #[actix_rt::test] async fn get_batch_filter_error() { let server = Server::new().await; let (response, code) = server.batches_filter("lol=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r#" { "message": "Unknown parameter `lol`: expected one of `limit`, `from`, `reverse`, `batchUids`, `uids`, `canceledBy`, `types`, `statuses`, `indexUids`, `afterEnqueuedAt`, `beforeEnqueuedAt`, `afterStartedAt`, `beforeStartedAt`, `afterFinishedAt`, `beforeFinishedAt`", "code": "bad_request", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#bad_request" } "#); let (response, code) = server.batches_filter("uids=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r#" { "message": "Invalid value in parameter `uids`: could not parse `pied` as a positive integer", "code": "invalid_task_uids", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_task_uids" } "#); let (response, code) = server.batches_filter("from=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r#" { "message": "Invalid value in parameter `from`: could not parse `pied` as a positive integer", "code": "invalid_task_from", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_task_from" } "#); let (response, code) = server.batches_filter("beforeStartedAt=pied").await; assert_eq!(code, 400, "{}", response); meili_snap::snapshot!(meili_snap::json_string!(response), @r#" { "message": "Invalid value in parameter `beforeStartedAt`: `pied` is an invalid date-time. It should follow the YYYY-MM-DD or RFC 3339 date-time format.", "code": "invalid_task_before_started_at", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_task_before_started_at" } "#); } #[actix_web::test] async fn test_summarized_document_addition_or_update() { let server = Server::new().await; let index = server.index("test"); index.add_documents(json!({ "id": 42, "content": "doggos & fluff" }), None).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "receivedDocuments": 1, "indexedDocuments": 1 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentAdditionOrUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.add_documents(json!({ "id": 42, "content": "doggos & fluff" }), Some("id")).await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "receivedDocuments": 1, "indexedDocuments": 1 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentAdditionOrUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_delete_documents_by_batch() { let server = Server::new().await; let index = server.index("test"); index.delete_batch(vec![1, 2, 3]).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "providedIds": 3, "deletedDocuments": 0 }, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(None).await; index.delete_batch(vec![42]).await; index.wait_task(2).await; let (batch, _) = index.get_batch(2).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 2, "details": { "providedIds": 1, "deletedDocuments": 0 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_delete_documents_by_filter() { let server = Server::new().await; let index = server.index("test"); index.delete_document_by_filter(json!({ "filter": "doggo = bernese" })).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "providedIds": 0, "deletedDocuments": 0, "originalFilter": "\"doggo = bernese\"" }, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(None).await; index.delete_document_by_filter(json!({ "filter": "doggo = bernese" })).await; index.wait_task(2).await; let (batch, _) = index.get_batch(2).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 2, "details": { "providedIds": 0, "deletedDocuments": 0, "originalFilter": "\"doggo = bernese\"" }, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.update_settings(json!({ "filterableAttributes": ["doggo"] })).await; index.delete_document_by_filter(json!({ "filter": "doggo = bernese" })).await; index.wait_task(4).await; let (batch, _) = index.get_batch(4).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 4, "details": { "providedIds": 0, "deletedDocuments": 0, "originalFilter": "\"doggo = bernese\"" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_delete_document_by_id() { let server = Server::new().await; let index = server.index("test"); index.delete_document(1).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "providedIds": 1, "deletedDocuments": 0 }, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(None).await; index.delete_document(42).await; index.wait_task(2).await; let (batch, _) = index.get_batch(2).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 2, "details": { "providedIds": 1, "deletedDocuments": 0 }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "documentDeletion": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_settings_update() { let server = Server::new().await; let index = server.index("test"); // here we should find my payload even in the failed batch. let (response, code) = index.update_settings(json!({ "rankingRules": ["custom"] })).await; meili_snap::snapshot!(code, @"400 Bad Request"); meili_snap::snapshot!(meili_snap::json_string!(response), @r###" { "message": "Invalid value at `.rankingRules[0]`: `custom` ranking rule is invalid. Valid ranking rules are words, typo, sort, proximity, attribute, exactness and custom ranking rules.", "code": "invalid_settings_ranking_rules", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#invalid_settings_ranking_rules" } "###); index.update_settings(json!({ "displayedAttributes": ["doggos", "name"], "filterableAttributes": ["age", "nb_paw_pads"], "sortableAttributes": ["iq"] })).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "displayedAttributes": [ "doggos", "name" ], "filterableAttributes": [ "age", "nb_paw_pads" ], "sortableAttributes": [ "iq" ] }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "settingsUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_index_creation() { let server = Server::new().await; let index = server.index("test"); index.create(None).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexCreation": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.create(Some("doggos")).await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "primaryKey": "doggos" }, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexCreation": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_index_deletion() { let server = Server::new().await; let index = server.index("test"); let (ret, _code) = index.delete().await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "failed", "type": "indexDeletion", "canceledBy": null, "details": { "deletedDocuments": 0 }, "error": { "message": "Index `test` not found.", "code": "index_not_found", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#index_not_found" }, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); // is the details correctly set when documents are actually deleted. // /!\ We need to wait for the document addition to be processed otherwise, if the test runs too slow, // both batches may get autobatched and the deleted documents count will be wrong. let (ret, _code) = index.add_documents(json!({ "id": 42, "content": "doggos & fluff" }), Some("id")).await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "succeeded", "type": "documentAdditionOrUpdate", "canceledBy": null, "details": { "receivedDocuments": 1, "indexedDocuments": 1 }, "error": null, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); let (ret, _code) = index.delete().await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "succeeded", "type": "indexDeletion", "canceledBy": null, "details": { "deletedDocuments": 1 }, "error": null, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); // What happens when you delete an index that doesn't exists. let (ret, _code) = index.delete().await; let batch = index.wait_task(ret.uid()).await; snapshot!(batch, @r###" { "uid": "[uid]", "batchUid": "[batch_uid]", "indexUid": "test", "status": "failed", "type": "indexDeletion", "canceledBy": null, "details": { "deletedDocuments": 0 }, "error": { "message": "Index `test` not found.", "code": "index_not_found", "type": "invalid_request", "link": "https://docs.meilisearch.com/errors#index_not_found" }, "duration": "[duration]", "enqueuedAt": "[date]", "startedAt": "[date]", "finishedAt": "[date]" } "###); } #[actix_web::test] async fn test_summarized_index_update() { let server = Server::new().await; let index = server.index("test"); // If the index doesn't exist yet, we should get errors with or without the primary key. index.update(None).await; index.wait_task(0).await; let (batch, _) = index.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": {}, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.update(Some("bones")).await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "primaryKey": "bones" }, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); // And run the same two tests once the index do exists. index.create(None).await; index.update(None).await; index.wait_task(3).await; let (batch, _) = index.get_batch(3).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 3, "details": {}, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); index.update(Some("bones")).await; index.wait_task(4).await; let (batch, _) = index.get_batch(4).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 4, "details": { "primaryKey": "bones" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexUpdate": 1 }, "indexUids": { "test": 1 } }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_index_swap() { let server = Server::new().await; server .index_swap(json!([ { "indexes": ["doggos", "cattos"] } ])) .await; server.wait_task(0).await; let (batch, _) = server.get_batch(0).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "swaps": [ { "indexes": [ "doggos", "cattos" ] } ] }, "stats": { "totalNbTasks": 1, "status": { "failed": 1 }, "types": { "indexSwap": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); server.index("doggos").create(None).await; server.index("cattos").create(None).await; server .index_swap(json!([ { "indexes": ["doggos", "cattos"] } ])) .await; server.wait_task(3).await; let (batch, _) = server.get_batch(3).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 3, "details": { "swaps": [ { "indexes": [ "doggos", "cattos" ] } ] }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "indexSwap": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_batch_cancelation() { let server = Server::new().await; let index = server.index("doggos"); // to avoid being flaky we're only going to cancel an already finished batch :( index.create(None).await; index.wait_task(0).await; server.cancel_tasks("uids=0").await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "matchedTasks": 1, "canceledTasks": 0, "originalFilter": "?uids=0" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "taskCancelation": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_batch_deletion() { let server = Server::new().await; let index = server.index("doggos"); // to avoid being flaky we're only going to delete an already finished batch :( index.create(None).await; index.wait_task(0).await; server.delete_tasks("uids=0").await; index.wait_task(1).await; let (batch, _) = index.get_batch(1).await; assert_json_snapshot!(batch, { ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 1, "details": { "matchedTasks": 1, "deletedTasks": 1, "originalFilter": "?uids=0" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "taskDeletion": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); } #[actix_web::test] async fn test_summarized_dump_creation() { let server = Server::new().await; server.create_dump().await; server.wait_task(0).await; let (batch, _) = server.get_batch(0).await; assert_json_snapshot!(batch, { ".details.dumpUid" => "[dumpUid]", ".duration" => "[duration]", ".enqueuedAt" => "[date]", ".startedAt" => "[date]", ".finishedAt" => "[date]" }, @r#" { "uid": 0, "details": { "dumpUid": "[dumpUid]" }, "stats": { "totalNbTasks": 1, "status": { "succeeded": 1 }, "types": { "dumpCreation": 1 }, "indexUids": {} }, "duration": "[duration]", "startedAt": "[date]", "finishedAt": "[date]" } "#); }