2024-11-19 15:05:34 +01:00
mod errors ;
use meili_snap ::insta ::assert_json_snapshot ;
use meili_snap ::snapshot ;
use crate ::common ::Server ;
use crate ::json ;
#[ actix_rt::test ]
async fn error_get_unexisting_batch_status ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
let ( response , code ) = index . get_batch ( 1 ) . await ;
let expected_response = json! ( {
2024-11-19 19:40:31 +01:00
" message " : " Batch `1` not found. " ,
2024-11-19 15:05:34 +01:00
" code " : " batch_not_found " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#batch_not_found "
} ) ;
assert_eq! ( response , expected_response ) ;
assert_eq! ( code , 404 ) ;
}
#[ actix_rt::test ]
async fn get_batch_status ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
2024-11-19 19:40:31 +01:00
let ( _response , code ) = index . get_batch ( 0 ) . await ;
2024-11-19 15:05:34 +01:00
assert_eq! ( code , 200 ) ;
}
#[ actix_rt::test ]
async fn list_batches ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
index
. add_documents ( serde_json ::from_str ( include_str! ( " ../assets/test_set.json " ) ) . unwrap ( ) , None )
. await ;
let ( response , code ) = index . list_batches ( ) . await ;
assert_eq! ( code , 200 ) ;
2024-11-20 17:48:26 +01:00
assert_eq! (
response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) ,
2 ,
" {} " ,
serde_json ::to_string_pretty ( & response ) . unwrap ( )
) ;
2024-11-19 15:05:34 +01:00
}
2024-11-20 13:29:52 +01:00
#[ actix_rt::test ]
async fn list_batches_pagination_and_reverse ( ) {
let server = Server ::new ( ) . await ;
// First of all we want to create a lot of batches very quickly. The fastest way is to delete a lot of unexisting indexes
let mut last_batch = None ;
for i in 0 .. 10 {
let index = server . index ( format! ( " test- {i} " ) ) ;
last_batch = Some ( index . create ( None ) . await . 0. uid ( ) ) ;
}
server . wait_task ( last_batch . unwrap ( ) ) . await ;
let ( response , code ) = server . batches_filter ( " limit=3 " ) . await ;
assert_eq! ( code , 200 ) ;
let results = response [ " results " ] . as_array ( ) . unwrap ( ) ;
let batch_ids : Vec < _ > = results . iter ( ) . map ( | ret | ret [ " uid " ] . as_u64 ( ) . unwrap ( ) ) . collect ( ) ;
snapshot! ( format! ( " {batch_ids:?} " ) , @ " [9, 8, 7] " ) ;
let ( response , code ) = server . batches_filter ( " limit=3&from=1 " ) . await ;
assert_eq! ( code , 200 ) ;
let results = response [ " results " ] . as_array ( ) . unwrap ( ) ;
let batch_ids : Vec < _ > = results . iter ( ) . map ( | ret | ret [ " uid " ] . as_u64 ( ) . unwrap ( ) ) . collect ( ) ;
snapshot! ( format! ( " {batch_ids:?} " ) , @ " [1, 0] " ) ;
// In reversed order
let ( response , code ) = server . batches_filter ( " limit=3&reverse=true " ) . await ;
assert_eq! ( code , 200 ) ;
let results = response [ " results " ] . as_array ( ) . unwrap ( ) ;
let batch_ids : Vec < _ > = results . iter ( ) . map ( | ret | ret [ " uid " ] . as_u64 ( ) . unwrap ( ) ) . collect ( ) ;
snapshot! ( format! ( " {batch_ids:?} " ) , @ " [0, 1, 2] " ) ;
let ( response , code ) = server . batches_filter ( " limit=3&from=8&reverse=true " ) . await ;
assert_eq! ( code , 200 ) ;
let results = response [ " results " ] . as_array ( ) . unwrap ( ) ;
let batch_ids : Vec < _ > = results . iter ( ) . map ( | ret | ret [ " uid " ] . as_u64 ( ) . unwrap ( ) ) . collect ( ) ;
snapshot! ( format! ( " {batch_ids:?} " ) , @ " [8, 9] " ) ;
}
2024-11-19 15:05:34 +01:00
#[ actix_rt::test ]
async fn list_batches_with_star_filters ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
let ( batch , _code ) = index . create ( None ) . await ;
index . wait_task ( batch . uid ( ) ) . await ;
index
. add_documents ( serde_json ::from_str ( include_str! ( " ../assets/test_set.json " ) ) . unwrap ( ) , None )
. await ;
let ( response , code ) = index . service . get ( " /batches?indexUids=test " ) . await ;
assert_eq! ( code , 200 ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
let ( response , code ) = index . service . get ( " /batches?indexUids=* " ) . await ;
assert_eq! ( code , 200 ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
let ( response , code ) = index . service . get ( " /batches?indexUids=*,pasteque " ) . await ;
assert_eq! ( code , 200 ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
let ( response , code ) = index . service . get ( " /batches?types=* " ) . await ;
assert_eq! ( code , 200 ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
let ( response , code ) =
index . service . get ( " /batches?types=*,documentAdditionOrUpdate&statuses=* " ) . await ;
assert_eq! ( code , 200 , " {:?} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
let ( response , code ) = index
. service
. get ( " /batches?types=*,documentAdditionOrUpdate&statuses=*,failed&indexUids=test " )
. await ;
assert_eq! ( code , 200 , " {:?} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
let ( response , code ) = index
. service
. get ( " /batches?types=*,documentAdditionOrUpdate&statuses=*,failed&indexUids=test,* " )
. await ;
assert_eq! ( code , 200 , " {:?} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
}
#[ actix_rt::test ]
async fn list_batches_status_filtered ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
index
. add_documents ( serde_json ::from_str ( include_str! ( " ../assets/test_set.json " ) ) . unwrap ( ) , None )
. await ;
let ( response , code ) = index . filtered_batches ( & [ ] , & [ " succeeded " ] , & [ ] ) . await ;
assert_eq! ( code , 200 , " {} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 1 ) ;
// We can't be sure that the update isn't already processed so we can't test this
// let (response, code) = index.filtered_batches(&[], &["processing"]).await;
// assert_eq!(code, 200, "{}", response);
// assert_eq!(response["results"].as_array().unwrap().len(), 1);
index . wait_task ( 1 ) . await ;
let ( response , code ) = index . filtered_batches ( & [ ] , & [ " succeeded " ] , & [ ] ) . await ;
assert_eq! ( code , 200 , " {} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
}
#[ actix_rt::test ]
async fn list_batches_type_filtered ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
index
. add_documents ( serde_json ::from_str ( include_str! ( " ../assets/test_set.json " ) ) . unwrap ( ) , None )
. await ;
let ( response , code ) = index . filtered_batches ( & [ " indexCreation " ] , & [ ] , & [ ] ) . await ;
assert_eq! ( code , 200 , " {} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 1 ) ;
let ( response , code ) =
index . filtered_batches ( & [ " indexCreation " , " documentAdditionOrUpdate " ] , & [ ] , & [ ] ) . await ;
assert_eq! ( code , 200 , " {} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
}
#[ actix_rt::test ]
async fn list_batches_invalid_canceled_by_filter ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
index
. add_documents ( serde_json ::from_str ( include_str! ( " ../assets/test_set.json " ) ) . unwrap ( ) , None )
. await ;
let ( response , code ) = index . filtered_batches ( & [ ] , & [ ] , & [ " 0 " ] ) . await ;
assert_eq! ( code , 200 , " {} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 0 ) ;
}
#[ actix_rt::test ]
async fn list_batches_status_and_type_filtered ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
index
. add_documents ( serde_json ::from_str ( include_str! ( " ../assets/test_set.json " ) ) . unwrap ( ) , None )
. await ;
let ( response , code ) = index . filtered_batches ( & [ " indexCreation " ] , & [ " failed " ] , & [ ] ) . await ;
assert_eq! ( code , 200 , " {} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 0 ) ;
let ( response , code ) = index
. filtered_batches (
& [ " indexCreation " , " documentAdditionOrUpdate " ] ,
& [ " succeeded " , " processing " , " enqueued " ] ,
& [ ] ,
)
. await ;
assert_eq! ( code , 200 , " {} " , response ) ;
assert_eq! ( response [ " results " ] . as_array ( ) . unwrap ( ) . len ( ) , 2 ) ;
}
#[ actix_rt::test ]
2024-12-02 12:02:16 +01:00
async fn list_batch_filter_error ( ) {
2024-11-19 15:05:34 +01:00
let server = Server ::new ( ) . await ;
let ( response , code ) = server . batches_filter ( " lol=pied " ) . await ;
assert_eq! ( code , 400 , " {} " , response ) ;
2024-11-20 10:54:38 +01:00
meili_snap ::snapshot! ( meili_snap ::json_string! ( response ) , @ r #"
2024-11-19 15:05:34 +01:00
{
2024-11-20 10:54:38 +01:00
" message " : " Unknown parameter `lol`: expected one of `limit`, `from`, `reverse`, `batchUids`, `uids`, `canceledBy`, `types`, `statuses`, `indexUids`, `afterEnqueuedAt`, `beforeEnqueuedAt`, `afterStartedAt`, `beforeStartedAt`, `afterFinishedAt`, `beforeFinishedAt` " ,
2024-11-19 15:05:34 +01:00
" code " : " bad_request " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#bad_request "
}
2024-11-20 10:54:38 +01:00
" #);
2024-11-19 15:05:34 +01:00
let ( response , code ) = server . batches_filter ( " uids=pied " ) . await ;
assert_eq! ( code , 400 , " {} " , response ) ;
meili_snap ::snapshot! ( meili_snap ::json_string! ( response ) , @ r #"
{
" message " : " Invalid value in parameter `uids`: could not parse `pied` as a positive integer " ,
" code " : " invalid_task_uids " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#invalid_task_uids "
}
" #);
let ( response , code ) = server . batches_filter ( " from=pied " ) . await ;
assert_eq! ( code , 400 , " {} " , response ) ;
meili_snap ::snapshot! ( meili_snap ::json_string! ( response ) , @ r #"
{
" message " : " Invalid value in parameter `from`: could not parse `pied` as a positive integer " ,
" code " : " invalid_task_from " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#invalid_task_from "
}
" #);
let ( response , code ) = server . batches_filter ( " beforeStartedAt=pied " ) . await ;
assert_eq! ( code , 400 , " {} " , response ) ;
meili_snap ::snapshot! ( meili_snap ::json_string! ( response ) , @ r #"
{
" message " : " Invalid value in parameter `beforeStartedAt`: `pied` is an invalid date-time. It should follow the YYYY-MM-DD or RFC 3339 date-time format. " ,
" code " : " invalid_task_before_started_at " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#invalid_task_before_started_at "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_document_addition_or_update ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . add_documents ( json! ( { " id " : 42 , " content " : " doggos & fluff " } ) , None ) . await ;
index . wait_task ( 0 ) . await ;
let ( batch , _ ) = index . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 18:19:41 +01:00
" details " : {
" receivedDocuments " : 1 ,
" indexedDocuments " : 1
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" documentAdditionOrUpdate " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . add_documents ( json! ( { " id " : 42 , " content " : " doggos & fluff " } ) , Some ( " id " ) ) . await ;
index . wait_task ( 1 ) . await ;
let ( batch , _ ) = index . get_batch ( 1 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 1 ,
2024-11-19 18:19:41 +01:00
" details " : {
" receivedDocuments " : 1 ,
" indexedDocuments " : 1
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" documentAdditionOrUpdate " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_delete_documents_by_batch ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . delete_batch ( vec! [ 1 , 2 , 3 ] ) . await ;
index . wait_task ( 0 ) . await ;
let ( batch , _ ) = index . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 19:40:31 +01:00
" details " : {
" providedIds " : 3 ,
" deletedDocuments " : 0
} ,
2024-11-19 18:19:41 +01:00
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" documentDeletion " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . create ( None ) . await ;
index . delete_batch ( vec! [ 42 ] ) . await ;
index . wait_task ( 2 ) . await ;
let ( batch , _ ) = index . get_batch ( 2 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 2 ,
2024-11-19 18:19:41 +01:00
" details " : {
" providedIds " : 1 ,
" deletedDocuments " : 0
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" documentDeletion " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_delete_documents_by_filter ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . delete_document_by_filter ( json! ( { " filter " : " doggo = bernese " } ) ) . await ;
index . wait_task ( 0 ) . await ;
let ( batch , _ ) = index . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 19:40:31 +01:00
" details " : {
" providedIds " : 0 ,
" deletedDocuments " : 0 ,
" originalFilter " : " \" doggo = bernese \" "
} ,
2024-11-19 18:19:41 +01:00
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" documentDeletion " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . create ( None ) . await ;
index . delete_document_by_filter ( json! ( { " filter " : " doggo = bernese " } ) ) . await ;
index . wait_task ( 2 ) . await ;
let ( batch , _ ) = index . get_batch ( 2 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 2 ,
2024-11-19 19:40:31 +01:00
" details " : {
" providedIds " : 0 ,
" deletedDocuments " : 0 ,
" originalFilter " : " \" doggo = bernese \" "
} ,
2024-11-19 18:19:41 +01:00
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" documentDeletion " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . update_settings ( json! ( { " filterableAttributes " : [ " doggo " ] } ) ) . await ;
index . delete_document_by_filter ( json! ( { " filter " : " doggo = bernese " } ) ) . await ;
index . wait_task ( 4 ) . await ;
let ( batch , _ ) = index . get_batch ( 4 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 4 ,
2024-11-19 18:19:41 +01:00
" details " : {
" providedIds " : 0 ,
" deletedDocuments " : 0 ,
" originalFilter " : " \" doggo = bernese \" "
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" documentDeletion " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_delete_document_by_id ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . delete_document ( 1 ) . await ;
index . wait_task ( 0 ) . await ;
let ( batch , _ ) = index . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 19:40:31 +01:00
" details " : {
" providedIds " : 1 ,
" deletedDocuments " : 0
} ,
2024-11-19 18:19:41 +01:00
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" documentDeletion " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . create ( None ) . await ;
index . delete_document ( 42 ) . await ;
index . wait_task ( 2 ) . await ;
let ( batch , _ ) = index . get_batch ( 2 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 2 ,
2024-11-19 18:19:41 +01:00
" details " : {
" providedIds " : 1 ,
" deletedDocuments " : 0
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" documentDeletion " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_settings_update ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
// here we should find my payload even in the failed batch.
let ( response , code ) = index . update_settings ( json! ( { " rankingRules " : [ " custom " ] } ) ) . await ;
meili_snap ::snapshot! ( code , @ " 400 Bad Request " ) ;
meili_snap ::snapshot! ( meili_snap ::json_string! ( response ) , @ r ###"
{
" message " : " Invalid value at `.rankingRules[0]`: `custom` ranking rule is invalid. Valid ranking rules are words, typo, sort, proximity, attribute, exactness and custom ranking rules. " ,
" code " : " invalid_settings_ranking_rules " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#invalid_settings_ranking_rules "
}
" ###);
index . update_settings ( json! ( { " displayedAttributes " : [ " doggos " , " name " ] , " filterableAttributes " : [ " age " , " nb_paw_pads " ] , " sortableAttributes " : [ " iq " ] } ) ) . await ;
index . wait_task ( 0 ) . await ;
let ( batch , _ ) = index . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 18:19:41 +01:00
" details " : {
" displayedAttributes " : [
" doggos " ,
" name "
] ,
" filterableAttributes " : [
" age " ,
" nb_paw_pads "
] ,
" sortableAttributes " : [
" iq "
]
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" settingsUpdate " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_index_creation ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
let ( batch , _ ) = index . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 18:19:41 +01:00
" details " : { } ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" indexCreation " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . create ( Some ( " doggos " ) ) . await ;
index . wait_task ( 1 ) . await ;
let ( batch , _ ) = index . get_batch ( 1 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 1 ,
2024-11-19 19:40:31 +01:00
" details " : {
" primaryKey " : " doggos "
} ,
2024-11-19 18:19:41 +01:00
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" indexCreation " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_index_deletion ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
let ( ret , _code ) = index . delete ( ) . await ;
let batch = index . wait_task ( ret . uid ( ) ) . await ;
snapshot! ( batch ,
@ r ###"
{
" uid " : " [uid] " ,
" batchUid " : " [batch_uid] " ,
" indexUid " : " test " ,
" status " : " failed " ,
" type " : " indexDeletion " ,
" canceledBy " : null ,
" details " : {
" deletedDocuments " : 0
} ,
" error " : {
" message " : " Index `test` not found. " ,
" code " : " index_not_found " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#index_not_found "
} ,
" duration " : " [duration] " ,
" enqueuedAt " : " [date] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" ###);
// is the details correctly set when documents are actually deleted.
// /!\ We need to wait for the document addition to be processed otherwise, if the test runs too slow,
// both batches may get autobatched and the deleted documents count will be wrong.
let ( ret , _code ) =
index . add_documents ( json! ( { " id " : 42 , " content " : " doggos & fluff " } ) , Some ( " id " ) ) . await ;
let batch = index . wait_task ( ret . uid ( ) ) . await ;
snapshot! ( batch ,
@ r ###"
{
" uid " : " [uid] " ,
" batchUid " : " [batch_uid] " ,
" indexUid " : " test " ,
" status " : " succeeded " ,
" type " : " documentAdditionOrUpdate " ,
" canceledBy " : null ,
" details " : {
" receivedDocuments " : 1 ,
" indexedDocuments " : 1
} ,
" error " : null ,
" duration " : " [duration] " ,
" enqueuedAt " : " [date] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" ###);
let ( ret , _code ) = index . delete ( ) . await ;
let batch = index . wait_task ( ret . uid ( ) ) . await ;
snapshot! ( batch ,
@ r ###"
{
" uid " : " [uid] " ,
" batchUid " : " [batch_uid] " ,
" indexUid " : " test " ,
" status " : " succeeded " ,
" type " : " indexDeletion " ,
" canceledBy " : null ,
" details " : {
" deletedDocuments " : 1
} ,
" error " : null ,
" duration " : " [duration] " ,
" enqueuedAt " : " [date] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" ###);
// What happens when you delete an index that doesn't exists.
let ( ret , _code ) = index . delete ( ) . await ;
let batch = index . wait_task ( ret . uid ( ) ) . await ;
snapshot! ( batch ,
@ r ###"
{
" uid " : " [uid] " ,
" batchUid " : " [batch_uid] " ,
" indexUid " : " test " ,
" status " : " failed " ,
" type " : " indexDeletion " ,
" canceledBy " : null ,
" details " : {
" deletedDocuments " : 0
} ,
" error " : {
" message " : " Index `test` not found. " ,
" code " : " index_not_found " ,
" type " : " invalid_request " ,
" link " : " https://docs.meilisearch.com/errors#index_not_found "
} ,
" duration " : " [duration] " ,
" enqueuedAt " : " [date] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" ###);
}
#[ actix_web::test ]
async fn test_summarized_index_update ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " test " ) ;
// If the index doesn't exist yet, we should get errors with or without the primary key.
index . update ( None ) . await ;
index . wait_task ( 0 ) . await ;
let ( batch , _ ) = index . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 18:19:41 +01:00
" details " : { } ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" indexUpdate " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . update ( Some ( " bones " ) ) . await ;
index . wait_task ( 1 ) . await ;
let ( batch , _ ) = index . get_batch ( 1 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 1 ,
2024-11-19 19:40:31 +01:00
" details " : {
" primaryKey " : " bones "
} ,
2024-11-19 18:19:41 +01:00
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" indexUpdate " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
// And run the same two tests once the index do exists.
index . create ( None ) . await ;
index . update ( None ) . await ;
index . wait_task ( 3 ) . await ;
let ( batch , _ ) = index . get_batch ( 3 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 3 ,
2024-11-19 18:19:41 +01:00
" details " : { } ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" indexUpdate " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
index . update ( Some ( " bones " ) ) . await ;
index . wait_task ( 4 ) . await ;
let ( batch , _ ) = index . get_batch ( 4 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 4 ,
2024-11-19 18:19:41 +01:00
" details " : {
" primaryKey " : " bones "
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" indexUpdate " : 1
} ,
" indexUids " : {
" test " : 1
}
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_index_swap ( ) {
let server = Server ::new ( ) . await ;
server
. index_swap ( json! ( [
{ " indexes " : [ " doggos " , " cattos " ] }
] ) )
. await ;
server . wait_task ( 0 ) . await ;
let ( batch , _ ) = server . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 19:40:31 +01:00
" details " : {
" swaps " : [
{
" indexes " : [
" doggos " ,
" cattos "
]
}
]
} ,
2024-11-19 18:19:41 +01:00
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" failed " : 1
} ,
" types " : {
" indexSwap " : 1
} ,
" indexUids " : { }
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
server . index ( " doggos " ) . create ( None ) . await ;
server . index ( " cattos " ) . create ( None ) . await ;
server
. index_swap ( json! ( [
{ " indexes " : [ " doggos " , " cattos " ] }
] ) )
. await ;
server . wait_task ( 3 ) . await ;
let ( batch , _ ) = server . get_batch ( 3 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 3 ,
2024-11-19 18:19:41 +01:00
" details " : {
" swaps " : [
{
" indexes " : [
" doggos " ,
" cattos "
]
}
]
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" indexSwap " : 1
} ,
" indexUids " : { }
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_batch_cancelation ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " doggos " ) ;
// to avoid being flaky we're only going to cancel an already finished batch :(
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
server . cancel_tasks ( " uids=0 " ) . await ;
index . wait_task ( 1 ) . await ;
let ( batch , _ ) = index . get_batch ( 1 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 1 ,
2024-11-19 18:19:41 +01:00
" details " : {
" matchedTasks " : 1 ,
" canceledTasks " : 0 ,
" originalFilter " : " ?uids=0 "
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" taskCancelation " : 1
} ,
" indexUids " : { }
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_batch_deletion ( ) {
let server = Server ::new ( ) . await ;
let index = server . index ( " doggos " ) ;
// to avoid being flaky we're only going to delete an already finished batch :(
index . create ( None ) . await ;
index . wait_task ( 0 ) . await ;
server . delete_tasks ( " uids=0 " ) . await ;
index . wait_task ( 1 ) . await ;
let ( batch , _ ) = index . get_batch ( 1 ) . await ;
assert_json_snapshot! ( batch ,
{ " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 1 ,
2024-11-19 18:19:41 +01:00
" details " : {
" matchedTasks " : 1 ,
" deletedTasks " : 1 ,
" originalFilter " : " ?uids=0 "
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" taskDeletion " : 1
} ,
" indexUids " : { }
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}
#[ actix_web::test ]
async fn test_summarized_dump_creation ( ) {
let server = Server ::new ( ) . await ;
server . create_dump ( ) . await ;
server . wait_task ( 0 ) . await ;
let ( batch , _ ) = server . get_batch ( 0 ) . await ;
assert_json_snapshot! ( batch ,
{ " .details.dumpUid " = > " [dumpUid] " , " .duration " = > " [duration] " , " .enqueuedAt " = > " [date] " , " .startedAt " = > " [date] " , " .finishedAt " = > " [date] " } ,
@ r #"
{
" uid " : 0 ,
2024-11-19 18:19:41 +01:00
" details " : {
" dumpUid " : " [dumpUid] "
} ,
" stats " : {
" totalNbTasks " : 1 ,
" status " : {
" succeeded " : 1
} ,
" types " : {
" dumpCreation " : 1
} ,
" indexUids " : { }
} ,
2024-11-19 15:05:34 +01:00
" duration " : " [duration] " ,
" startedAt " : " [date] " ,
" finishedAt " : " [date] "
}
" #);
}