mirror of
https://github.com/meilisearch/MeiliSearch
synced 2024-11-30 08:44:27 +01:00
364 lines
11 KiB
Rust
364 lines
11 KiB
Rust
|
use assert_json_diff::{assert_json_eq, assert_json_include};
|
||
|
use meilisearch_http::helpers::compression;
|
||
|
use serde_json::{json, Value};
|
||
|
use std::fs::File;
|
||
|
use std::path::Path;
|
||
|
use std::thread;
|
||
|
use std::time::Duration;
|
||
|
use tempfile::TempDir;
|
||
|
|
||
|
#[macro_use] mod common;
|
||
|
|
||
|
async fn trigger_and_wait_backup(server: &mut common::Server) -> String {
|
||
|
let (value, status_code) = server.trigger_backup().await;
|
||
|
|
||
|
assert_eq!(status_code, 202);
|
||
|
|
||
|
let backup_uid = value["uid"].as_str().unwrap().to_string();
|
||
|
|
||
|
for _ in 0..20 as u8 {
|
||
|
let (value, status_code) = server.get_backup_status(&backup_uid).await;
|
||
|
|
||
|
assert_eq!(status_code, 200);
|
||
|
assert_ne!(value["status"].as_str(), Some("backup_process_failed"));
|
||
|
|
||
|
if value["status"].as_str() == Some("done") { return backup_uid }
|
||
|
thread::sleep(Duration::from_millis(100));
|
||
|
}
|
||
|
|
||
|
unreachable!("backup creation runned out of time")
|
||
|
}
|
||
|
|
||
|
fn current_db_version() -> (String, String, String) {
|
||
|
let current_version_major = env!("CARGO_PKG_VERSION_MAJOR").to_string();
|
||
|
let current_version_minor = env!("CARGO_PKG_VERSION_MINOR").to_string();
|
||
|
let current_version_patch = env!("CARGO_PKG_VERSION_PATCH").to_string();
|
||
|
|
||
|
(current_version_major, current_version_minor, current_version_patch)
|
||
|
}
|
||
|
|
||
|
fn current_backup_version() -> String {
|
||
|
"V1".into()
|
||
|
}
|
||
|
|
||
|
fn read_all_jsonline<R: std::io::Read>(r: R) -> Value {
|
||
|
let deserializer = serde_json::Deserializer::from_reader(r);
|
||
|
let iterator = deserializer.into_iter::<serde_json::Value>();
|
||
|
|
||
|
json!(iterator.map(|v| v.unwrap()).collect::<Vec<Value>>())
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn trigger_backup_should_return_ok() {
|
||
|
let server = common::Server::test_server().await;
|
||
|
|
||
|
let (_, status_code) = server.trigger_backup().await;
|
||
|
|
||
|
assert_eq!(status_code, 202);
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn trigger_backup_twice_should_return_conflict() {
|
||
|
let server = common::Server::test_server().await;
|
||
|
|
||
|
let expected = json!({
|
||
|
"message": "Another backup is already in progress",
|
||
|
"errorCode": "backup_already_in_progress",
|
||
|
"errorType": "invalid_request_error",
|
||
|
"errorLink": "https://docs.meilisearch.com/errors#backup_already_in_progress"
|
||
|
});
|
||
|
|
||
|
let (_, status_code) = server.trigger_backup().await;
|
||
|
|
||
|
assert_eq!(status_code, 202);
|
||
|
|
||
|
let (value, status_code) = server.trigger_backup().await;
|
||
|
|
||
|
|
||
|
assert_json_eq!(expected.clone(), value.clone(), ordered: false);
|
||
|
assert_eq!(status_code, 409);
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn trigger_backup_concurently_should_return_conflict() {
|
||
|
let server = common::Server::test_server().await;
|
||
|
|
||
|
let expected = json!({
|
||
|
"message": "Another backup is already in progress",
|
||
|
"errorCode": "backup_already_in_progress",
|
||
|
"errorType": "invalid_request_error",
|
||
|
"errorLink": "https://docs.meilisearch.com/errors#backup_already_in_progress"
|
||
|
});
|
||
|
|
||
|
let ((_value_1, _status_code_1), (value_2, status_code_2)) = futures::join!(server.trigger_backup(), server.trigger_backup());
|
||
|
|
||
|
assert_json_eq!(expected.clone(), value_2.clone(), ordered: false);
|
||
|
assert_eq!(status_code_2, 409);
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn get_backup_status_early_should_return_processing() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
|
||
|
|
||
|
let (value, status_code) = server.trigger_backup().await;
|
||
|
|
||
|
assert_eq!(status_code, 202);
|
||
|
|
||
|
let backup_uid = value["uid"].as_str().unwrap().to_string();
|
||
|
|
||
|
let (value, status_code) = server.get_backup_status(&backup_uid).await;
|
||
|
|
||
|
let expected = json!({
|
||
|
"uid": backup_uid,
|
||
|
"status": "processing"
|
||
|
});
|
||
|
|
||
|
assert_eq!(status_code, 200);
|
||
|
|
||
|
assert_json_eq!(expected.clone(), value.clone(), ordered: false);
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn get_backup_status_should_return_done() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
|
||
|
let (value, status_code) = server.trigger_backup().await;
|
||
|
|
||
|
assert_eq!(status_code, 202);
|
||
|
|
||
|
println!("{:?}", value);
|
||
|
let backup_uid = value["uid"].as_str().unwrap().to_string();
|
||
|
|
||
|
let expected = json!({
|
||
|
"uid": backup_uid.clone(),
|
||
|
"status": "done"
|
||
|
});
|
||
|
|
||
|
thread::sleep(Duration::from_secs(1)); // wait backup until process end
|
||
|
|
||
|
let (value, status_code) = server.get_backup_status(&backup_uid).await;
|
||
|
|
||
|
assert_eq!(status_code, 200);
|
||
|
|
||
|
assert_json_eq!(expected.clone(), value.clone(), ordered: false);
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn backup_metadata_should_be_valid() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
let body = json!({
|
||
|
"uid": "test2",
|
||
|
"primaryKey": "test2_id",
|
||
|
});
|
||
|
|
||
|
server.create_index(body).await;
|
||
|
|
||
|
let uid = trigger_and_wait_backup(&mut server).await;
|
||
|
|
||
|
let backup_folder = Path::new(&server.data().backup_folder);
|
||
|
let tmp_dir = TempDir::new().unwrap();
|
||
|
let tmp_dir_path = tmp_dir.path();
|
||
|
|
||
|
compression::from_tar_gz(&backup_folder.join(&format!("{}.tar.gz", uid)), tmp_dir_path).unwrap();
|
||
|
|
||
|
let file = File::open(tmp_dir_path.join("metadata.json")).unwrap();
|
||
|
let mut metadata: serde_json::Value = serde_json::from_reader(file).unwrap();
|
||
|
|
||
|
// fields are randomly ordered
|
||
|
metadata.get_mut("indexes").unwrap()
|
||
|
.as_array_mut().unwrap()
|
||
|
.sort_by(|a, b|
|
||
|
a.get("uid").unwrap().as_str().cmp(&b.get("uid").unwrap().as_str())
|
||
|
);
|
||
|
|
||
|
let (major, minor, patch) = current_db_version();
|
||
|
|
||
|
let expected = json!({
|
||
|
"indexes": [{
|
||
|
"uid": "test",
|
||
|
"primaryKey": "id",
|
||
|
}, {
|
||
|
"uid": "test2",
|
||
|
"primaryKey": "test2_id",
|
||
|
}
|
||
|
],
|
||
|
"dbVersion": format!("{}.{}.{}", major, minor, patch),
|
||
|
"backupVersion": current_backup_version()
|
||
|
});
|
||
|
|
||
|
assert_json_include!(expected: expected.clone(), actual: metadata.clone());
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn backup_gzip_should_have_been_created() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
|
||
|
let backup_uid = trigger_and_wait_backup(&mut server).await;
|
||
|
let backup_folder = Path::new(&server.data().backup_folder);
|
||
|
|
||
|
let compressed_path = backup_folder.join(format!("{}.tar.gz", backup_uid));
|
||
|
assert!(File::open(compressed_path).is_ok());
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn backup_index_settings_should_be_valid() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
let expected = json!({
|
||
|
"rankingRules": [
|
||
|
"typo",
|
||
|
"words",
|
||
|
"proximity",
|
||
|
"attribute",
|
||
|
"wordsPosition",
|
||
|
"exactness"
|
||
|
],
|
||
|
"distinctAttribute": "email",
|
||
|
"searchableAttributes": [
|
||
|
"balance",
|
||
|
"picture",
|
||
|
"age",
|
||
|
"color",
|
||
|
"name",
|
||
|
"gender",
|
||
|
"email",
|
||
|
"phone",
|
||
|
"address",
|
||
|
"about",
|
||
|
"registered",
|
||
|
"latitude",
|
||
|
"longitude",
|
||
|
"tags"
|
||
|
],
|
||
|
"displayedAttributes": [
|
||
|
"id",
|
||
|
"isActive",
|
||
|
"balance",
|
||
|
"picture",
|
||
|
"age",
|
||
|
"color",
|
||
|
"name",
|
||
|
"gender",
|
||
|
"email",
|
||
|
"phone",
|
||
|
"address",
|
||
|
"about",
|
||
|
"registered",
|
||
|
"latitude",
|
||
|
"longitude",
|
||
|
"tags"
|
||
|
],
|
||
|
"stopWords": [
|
||
|
"in",
|
||
|
"ad"
|
||
|
],
|
||
|
"synonyms": {
|
||
|
"wolverine": ["xmen", "logan"],
|
||
|
"logan": ["wolverine", "xmen"]
|
||
|
},
|
||
|
"attributesForFaceting": [
|
||
|
"gender",
|
||
|
"color",
|
||
|
"tags"
|
||
|
]
|
||
|
});
|
||
|
|
||
|
server.update_all_settings(expected.clone()).await;
|
||
|
|
||
|
let uid = trigger_and_wait_backup(&mut server).await;
|
||
|
|
||
|
let backup_folder = Path::new(&server.data().backup_folder);
|
||
|
let tmp_dir = TempDir::new().unwrap();
|
||
|
let tmp_dir_path = tmp_dir.path();
|
||
|
|
||
|
compression::from_tar_gz(&backup_folder.join(&format!("{}.tar.gz", uid)), tmp_dir_path).unwrap();
|
||
|
|
||
|
let file = File::open(tmp_dir_path.join("test").join("settings.json")).unwrap();
|
||
|
let settings: serde_json::Value = serde_json::from_reader(file).unwrap();
|
||
|
|
||
|
assert_json_eq!(expected.clone(), settings.clone(), ordered: false);
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn backup_index_documents_should_be_valid() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
let dataset = include_bytes!("assets/backups/v1/test/documents.jsonl");
|
||
|
let mut slice: &[u8] = dataset;
|
||
|
|
||
|
let expected: Value = read_all_jsonline(&mut slice);
|
||
|
|
||
|
let uid = trigger_and_wait_backup(&mut server).await;
|
||
|
|
||
|
let backup_folder = Path::new(&server.data().backup_folder);
|
||
|
let tmp_dir = TempDir::new().unwrap();
|
||
|
let tmp_dir_path = tmp_dir.path();
|
||
|
|
||
|
compression::from_tar_gz(&backup_folder.join(&format!("{}.tar.gz", uid)), tmp_dir_path).unwrap();
|
||
|
|
||
|
let file = File::open(tmp_dir_path.join("test").join("documents.jsonl")).unwrap();
|
||
|
let documents = read_all_jsonline(file);
|
||
|
|
||
|
assert_json_eq!(expected.clone(), documents.clone(), ordered: false);
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn backup_index_updates_should_be_valid() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
let dataset = include_bytes!("assets/backups/v1/test/updates.jsonl");
|
||
|
let mut slice: &[u8] = dataset;
|
||
|
|
||
|
let expected: Value = read_all_jsonline(&mut slice);
|
||
|
|
||
|
let uid = trigger_and_wait_backup(&mut server).await;
|
||
|
|
||
|
let backup_folder = Path::new(&server.data().backup_folder);
|
||
|
let tmp_dir = TempDir::new().unwrap();
|
||
|
let tmp_dir_path = tmp_dir.path();
|
||
|
|
||
|
compression::from_tar_gz(&backup_folder.join(&format!("{}.tar.gz", uid)), tmp_dir_path).unwrap();
|
||
|
|
||
|
let file = File::open(tmp_dir_path.join("test").join("updates.jsonl")).unwrap();
|
||
|
let mut updates = read_all_jsonline(file);
|
||
|
|
||
|
|
||
|
// hotfix until #943 is fixed (https://github.com/meilisearch/MeiliSearch/issues/943)
|
||
|
updates.as_array_mut().unwrap()
|
||
|
.get_mut(0).unwrap()
|
||
|
.get_mut("type").unwrap()
|
||
|
.get_mut("settings").unwrap()
|
||
|
.get_mut("displayed_attributes").unwrap()
|
||
|
.get_mut("Update").unwrap()
|
||
|
.as_array_mut().unwrap().sort_by(|a, b| a.as_str().cmp(&b.as_str()));
|
||
|
|
||
|
eprintln!("{}\n", updates.to_string());
|
||
|
eprintln!("{}", expected.to_string());
|
||
|
assert_json_include!(expected: expected.clone(), actual: updates.clone());
|
||
|
}
|
||
|
|
||
|
#[actix_rt::test]
|
||
|
#[ignore]
|
||
|
async fn get_unexisting_backup_status_should_return_not_found() {
|
||
|
let mut server = common::Server::test_server().await;
|
||
|
|
||
|
let (_, status_code) = server.get_backup_status("4242").await;
|
||
|
|
||
|
assert_eq!(status_code, 404);
|
||
|
}
|