mirror of
https://github.com/meilisearch/MeiliSearch
synced 2024-11-26 23:04:26 +01:00
730 lines
21 KiB
Rust
730 lines
21 KiB
Rust
use actix_web::http::header::ACCEPT_ENCODING;
|
|
use actix_web::test;
|
|
use meili_snap::*;
|
|
use urlencoding::encode as urlencode;
|
|
|
|
use crate::common::encoder::Encoder;
|
|
use crate::common::{shared_does_not_exists_index, shared_empty_index, shared_index_with_test_set, GetAllDocumentsOptions, Server, Value};
|
|
use crate::json;
|
|
|
|
// TODO: partial test since we are testing error, amd error is not yet fully implemented in
|
|
// transplant
|
|
#[actix_rt::test]
|
|
async fn get_unexisting_index_single_document() {
|
|
let (_response, code) = shared_does_not_exists_index().await.get_document(1, None).await;
|
|
assert_eq!(code, 404);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn error_get_unexisting_document() {
|
|
let server = Server::new_shared();
|
|
let index = server.unique_index();
|
|
let (task, _code) = index.create(None).await;
|
|
index.wait_task(task.uid()).await.succeeded();
|
|
|
|
let (response, code) = index.get_document(1, None).await;
|
|
|
|
let expected_response = json!({
|
|
"message": "Document `1` not found.",
|
|
"code": "document_not_found",
|
|
"type": "invalid_request",
|
|
"link": "https://docs.meilisearch.com/errors#document_not_found"
|
|
});
|
|
|
|
assert_eq!(response, expected_response);
|
|
assert_eq!(code, 404);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_document() {
|
|
let server = Server::new_shared();
|
|
let index = server.unique_index();
|
|
let (task, _code) = index.create(None).await;
|
|
index.wait_task(task.uid()).await.succeeded();
|
|
let documents = json!([
|
|
{
|
|
"id": 0,
|
|
"nested": { "content": "foobar" },
|
|
}
|
|
]);
|
|
let (task, code) = index.add_documents(documents, None).await;
|
|
assert_eq!(code, 202);
|
|
index.wait_task(task.uid()).await.succeeded();
|
|
let (response, code) = index.get_document(0, None).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(
|
|
response,
|
|
json!({
|
|
"id": 0,
|
|
"nested": { "content": "foobar" },
|
|
})
|
|
);
|
|
|
|
let (response, code) = index.get_document(0, Some(json!({ "fields": ["id"] }))).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(
|
|
response,
|
|
json!({
|
|
"id": 0,
|
|
})
|
|
);
|
|
|
|
let (response, code) =
|
|
index.get_document(0, Some(json!({ "fields": ["nested.content"] }))).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(
|
|
response,
|
|
json!({
|
|
"nested": { "content": "foobar" },
|
|
})
|
|
);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn error_get_unexisting_index_all_documents() {
|
|
let index = shared_does_not_exists_index().await;
|
|
let (response, code) = index.get_all_documents(GetAllDocumentsOptions::default()).await;
|
|
|
|
let expected_response = json!({
|
|
"message": "Index `DOES_NOT_EXISTS` not found.",
|
|
"code": "index_not_found",
|
|
"type": "invalid_request",
|
|
"link": "https://docs.meilisearch.com/errors#index_not_found"
|
|
});
|
|
|
|
assert_eq!(response, expected_response);
|
|
assert_eq!(code, 404);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_no_document() {
|
|
let index = shared_empty_index().await;
|
|
|
|
let (response, code) = index.get_all_documents(GetAllDocumentsOptions::default()).await;
|
|
assert_eq!(code, 200);
|
|
assert!(response["results"].as_array().unwrap().is_empty());
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_all_documents_no_options() {
|
|
let index = shared_index_with_test_set().await;
|
|
|
|
let (response, code) = index.get_all_documents(GetAllDocumentsOptions::default()).await;
|
|
assert_eq!(code, 200);
|
|
let results = response["results"].as_array().unwrap();
|
|
assert_eq!(results.len(), 20);
|
|
let first = json!({
|
|
"id":0,
|
|
"isActive":false,
|
|
"balance":"$2,668.55",
|
|
"picture":"http://placehold.it/32x32",
|
|
"age":36,
|
|
"color":"Green",
|
|
"name":"Lucas Hess",
|
|
"gender":"male",
|
|
"email":"lucashess@chorizon.com",
|
|
"phone":"+1 (998) 478-2597",
|
|
"address":"412 Losee Terrace, Blairstown, Georgia, 2825",
|
|
"about":"Mollit ad in exercitation quis. Anim est ut consequat fugiat duis magna aliquip velit nisi. Commodo eiusmod est consequat proident consectetur aliqua enim fugiat. Aliqua adipisicing laboris elit proident enim veniam laboris mollit. Incididunt fugiat minim ad nostrud deserunt tempor in. Id irure officia labore qui est labore nulla nisi. Magna sit quis tempor esse consectetur amet labore duis aliqua consequat.\r\n",
|
|
"registered":"2016-06-21T09:30:25 -02:00",
|
|
"latitude":-44.174957,
|
|
"longitude":-145.725388,
|
|
"tags":["bug"
|
|
,"bug"]});
|
|
assert_eq!(first, results[0]);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_all_documents_no_options_with_response_compression() {
|
|
let index = shared_index_with_test_set().await;
|
|
|
|
let app = Server::new_shared().init_web_app().await;
|
|
let req = test::TestRequest::get()
|
|
.uri(&format!("/indexes/{}/documents?", urlencode(&index.uid)))
|
|
.insert_header((ACCEPT_ENCODING, "gzip"))
|
|
.to_request();
|
|
|
|
let res = test::call_service(&app, req).await;
|
|
|
|
assert_eq!(res.status(), 200);
|
|
|
|
let bytes = test::read_body(res).await;
|
|
let decoded = Encoder::Gzip.decode(bytes);
|
|
let parsed_response =
|
|
serde_json::from_slice::<Value>(decoded.into().as_ref()).expect("Expecting valid json");
|
|
|
|
let arr = parsed_response["results"].as_array().unwrap();
|
|
assert_eq!(arr.len(), 20);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn test_get_all_documents_limit() {
|
|
let index = shared_index_with_test_set().await;
|
|
|
|
let (response, code) = index
|
|
.get_all_documents(GetAllDocumentsOptions { limit: Some(5), ..Default::default() })
|
|
.await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 5);
|
|
assert_eq!(response["results"][0]["id"], json!(0));
|
|
assert_eq!(response["offset"], json!(0));
|
|
assert_eq!(response["limit"], json!(5));
|
|
assert_eq!(response["total"], json!(77));
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn test_get_all_documents_offset() {
|
|
let index = shared_index_with_test_set().await;
|
|
|
|
let (response, code) = index
|
|
.get_all_documents(GetAllDocumentsOptions { offset: Some(5), ..Default::default() })
|
|
.await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
assert_eq!(response["results"][0]["id"], json!(5));
|
|
assert_eq!(response["offset"], json!(5));
|
|
assert_eq!(response["limit"], json!(20));
|
|
assert_eq!(response["total"], json!(77));
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn test_get_all_documents_attributes_to_retrieve() {
|
|
let index = shared_index_with_test_set().await;
|
|
|
|
|
|
let (response, code) = index
|
|
.get_all_documents(GetAllDocumentsOptions {
|
|
fields: Some(vec!["name"]),
|
|
..Default::default()
|
|
})
|
|
.await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
for results in response["results"].as_array().unwrap() {
|
|
assert_eq!(results.as_object().unwrap().keys().count(), 1);
|
|
assert!(results["name"] != json!(null));
|
|
}
|
|
assert_eq!(response["offset"], json!(0));
|
|
assert_eq!(response["limit"], json!(20));
|
|
assert_eq!(response["total"], json!(77));
|
|
|
|
let (response, code) = index.get_all_documents_raw("?fields=").await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
for results in response["results"].as_array().unwrap() {
|
|
assert_eq!(results.as_object().unwrap().keys().count(), 0);
|
|
}
|
|
assert_eq!(response["offset"], json!(0));
|
|
assert_eq!(response["limit"], json!(20));
|
|
assert_eq!(response["total"], json!(77));
|
|
|
|
let (response, code) = index
|
|
.get_all_documents(GetAllDocumentsOptions {
|
|
fields: Some(vec!["wrong"]),
|
|
..Default::default()
|
|
})
|
|
.await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
for results in response["results"].as_array().unwrap() {
|
|
assert_eq!(results.as_object().unwrap().keys().count(), 0);
|
|
}
|
|
assert_eq!(response["offset"], json!(0));
|
|
assert_eq!(response["limit"], json!(20));
|
|
assert_eq!(response["total"], json!(77));
|
|
|
|
let (response, code) = index
|
|
.get_all_documents(GetAllDocumentsOptions {
|
|
fields: Some(vec!["name", "tags"]),
|
|
..Default::default()
|
|
})
|
|
.await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
for results in response["results"].as_array().unwrap() {
|
|
assert_eq!(results.as_object().unwrap().keys().count(), 2);
|
|
assert!(results["name"] != json!(null));
|
|
assert!(results["tags"] != json!(null));
|
|
}
|
|
|
|
let (response, code) = index
|
|
.get_all_documents(GetAllDocumentsOptions { fields: Some(vec!["*"]), ..Default::default() })
|
|
.await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
for results in response["results"].as_array().unwrap() {
|
|
assert_eq!(results.as_object().unwrap().keys().count(), 16);
|
|
}
|
|
|
|
let (response, code) = index
|
|
.get_all_documents(GetAllDocumentsOptions {
|
|
fields: Some(vec!["*", "wrong"]),
|
|
..Default::default()
|
|
})
|
|
.await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
for results in response["results"].as_array().unwrap() {
|
|
assert_eq!(results.as_object().unwrap().keys().count(), 16);
|
|
}
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_document_s_nested_attributes_to_retrieve() {
|
|
let server = Server::new_shared();
|
|
let index = server.unique_index();
|
|
let (task, _code) = index.create(None).await;
|
|
index.wait_task(task.uid()).await.succeeded();
|
|
|
|
let documents = json!([
|
|
{
|
|
"id": 0,
|
|
"content.truc": "foobar",
|
|
},
|
|
{
|
|
"id": 1,
|
|
"content": {
|
|
"truc": "foobar",
|
|
"machin": "bidule",
|
|
},
|
|
},
|
|
]);
|
|
let (task, code) = index.add_documents(documents, None).await;
|
|
assert_eq!(code, 202);
|
|
index.wait_task(task.uid()).await.succeeded();
|
|
|
|
let (response, code) = index.get_document(0, Some(json!({ "fields": ["content"] }))).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response, json!({}));
|
|
let (response, code) = index.get_document(1, Some(json!({ "fields": ["content"] }))).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(
|
|
response,
|
|
json!({
|
|
"content": {
|
|
"truc": "foobar",
|
|
"machin": "bidule",
|
|
},
|
|
})
|
|
);
|
|
|
|
let (response, code) = index.get_document(0, Some(json!({ "fields": ["content.truc"] }))).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(
|
|
response,
|
|
json!({
|
|
"content.truc": "foobar",
|
|
})
|
|
);
|
|
let (response, code) = index.get_document(1, Some(json!({ "fields": ["content.truc"] }))).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(
|
|
response,
|
|
json!({
|
|
"content": {
|
|
"truc": "foobar",
|
|
},
|
|
})
|
|
);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_documents_displayed_attributes_is_ignored() {
|
|
let server = Server::new_shared();
|
|
let index = server.unique_index();
|
|
index.load_test_set().await;
|
|
index.update_settings(json!({"displayedAttributes": ["gender"]})).await;
|
|
|
|
let (response, code) = index.get_all_documents(GetAllDocumentsOptions::default()).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response["results"].as_array().unwrap().len(), 20);
|
|
assert_eq!(response["results"][0].as_object().unwrap().keys().count(), 16);
|
|
assert!(response["results"][0]["gender"] != json!(null));
|
|
|
|
assert_eq!(response["offset"], json!(0));
|
|
assert_eq!(response["limit"], json!(20));
|
|
assert_eq!(response["total"], json!(77));
|
|
|
|
let (response, code) = index.get_document(0, None).await;
|
|
assert_eq!(code, 200);
|
|
assert_eq!(response.as_object().unwrap().keys().count(), 16);
|
|
assert!(response.as_object().unwrap().get("gender").is_some());
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_document_by_filter() {
|
|
let server = Server::new_shared();
|
|
let index = server.unique_index();
|
|
index.update_settings_filterable_attributes(json!(["color"])).await;
|
|
let (task, _code) = index
|
|
.add_documents(
|
|
json!([
|
|
{ "id": 0, "color": "red" },
|
|
{ "id": 1, "color": "blue" },
|
|
{ "id": 2, "color": "blue" },
|
|
{ "id": 3 },
|
|
]),
|
|
Some("id"),
|
|
)
|
|
.await;
|
|
index.wait_task(task.uid()).await.succeeded();
|
|
|
|
let (response, code) = index.get_document_by_filter(json!({})).await;
|
|
let (response2, code2) = index.get_all_documents_raw("").await;
|
|
snapshot!(code, @"200 OK");
|
|
snapshot!(json_string!(response, { ".enqueuedAt" => "[date]" }), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"id": 0,
|
|
"color": "red"
|
|
},
|
|
{
|
|
"id": 1,
|
|
"color": "blue"
|
|
},
|
|
{
|
|
"id": 2,
|
|
"color": "blue"
|
|
},
|
|
{
|
|
"id": 3
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 4
|
|
}
|
|
"###);
|
|
assert_eq!(code, code2);
|
|
assert_eq!(response, response2);
|
|
|
|
let (response, code) = index.get_document_by_filter(json!({ "filter": "color = blue" })).await;
|
|
let (response2, code2) = index.get_all_documents_raw("?filter=color=blue").await;
|
|
snapshot!(code, @"200 OK");
|
|
snapshot!(json_string!(response, { ".enqueuedAt" => "[date]" }), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"id": 1,
|
|
"color": "blue"
|
|
},
|
|
{
|
|
"id": 2,
|
|
"color": "blue"
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 2
|
|
}
|
|
"###);
|
|
assert_eq!(code, code2);
|
|
assert_eq!(response, response2);
|
|
|
|
let (response, code) = index
|
|
.get_document_by_filter(json!({ "offset": 1, "limit": 1, "filter": "color != blue" }))
|
|
.await;
|
|
let (response2, code2) =
|
|
index.get_all_documents_raw("?filter=color!=blue&offset=1&limit=1").await;
|
|
snapshot!(code, @"200 OK");
|
|
snapshot!(json_string!(response, { ".enqueuedAt" => "[date]" }), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"id": 3
|
|
}
|
|
],
|
|
"offset": 1,
|
|
"limit": 1,
|
|
"total": 2
|
|
}
|
|
"###);
|
|
assert_eq!(code, code2);
|
|
assert_eq!(response, response2);
|
|
|
|
let (response, code) = index
|
|
.get_document_by_filter(
|
|
json!({ "limit": 1, "filter": "color != blue", "fields": ["color"] }),
|
|
)
|
|
.await;
|
|
let (response2, code2) =
|
|
index.get_all_documents_raw("?limit=1&filter=color!=blue&fields=color").await;
|
|
snapshot!(code, @"200 OK");
|
|
snapshot!(json_string!(response, { ".enqueuedAt" => "[date]" }), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"color": "red"
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 1,
|
|
"total": 2
|
|
}
|
|
"###);
|
|
assert_eq!(code, code2);
|
|
assert_eq!(response, response2);
|
|
|
|
// Now testing more complex filter that the get route can't represent
|
|
|
|
let (response, code) =
|
|
index.get_document_by_filter(json!({ "filter": [["color = blue", "color = red"]] })).await;
|
|
snapshot!(code, @"200 OK");
|
|
snapshot!(json_string!(response, { ".enqueuedAt" => "[date]" }), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"id": 0,
|
|
"color": "red"
|
|
},
|
|
{
|
|
"id": 1,
|
|
"color": "blue"
|
|
},
|
|
{
|
|
"id": 2,
|
|
"color": "blue"
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 3
|
|
}
|
|
"###);
|
|
|
|
let (response, code) = index
|
|
.get_document_by_filter(json!({ "filter": [["color != blue"], "color EXISTS"] }))
|
|
.await;
|
|
snapshot!(code, @"200 OK");
|
|
snapshot!(json_string!(response, { ".enqueuedAt" => "[date]" }), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"id": 0,
|
|
"color": "red"
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 1
|
|
}
|
|
"###);
|
|
}
|
|
|
|
#[actix_rt::test]
|
|
async fn get_document_with_vectors() {
|
|
let server = Server::new().await;
|
|
let index = server.index("doggo");
|
|
let (value, code) = server.set_features(json!({"vectorStore": true})).await;
|
|
snapshot!(code, @"200 OK");
|
|
snapshot!(value, @r###"
|
|
{
|
|
"vectorStore": true,
|
|
"metrics": false,
|
|
"logsRoute": false,
|
|
"editDocumentsByFunction": false,
|
|
"containsFilter": false
|
|
}
|
|
"###);
|
|
|
|
let (response, code) = index
|
|
.update_settings(json!({
|
|
"embedders": {
|
|
"manual": {
|
|
"source": "userProvided",
|
|
"dimensions": 3,
|
|
}
|
|
},
|
|
}))
|
|
.await;
|
|
snapshot!(code, @"202 Accepted");
|
|
server.wait_task(response.uid()).await.succeeded();
|
|
|
|
let documents = json!([
|
|
{"id": 0, "name": "kefir", "_vectors": { "manual": [0, 0, 0] }},
|
|
{"id": 1, "name": "echo", "_vectors": { "manual": null }},
|
|
]);
|
|
let (value, code) = index.add_documents(documents, None).await;
|
|
snapshot!(code, @"202 Accepted");
|
|
index.wait_task(value.uid()).await.succeeded();
|
|
|
|
// by default you shouldn't see the `_vectors` object
|
|
let (documents, _code) = index.get_all_documents(Default::default()).await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"id": 0,
|
|
"name": "kefir"
|
|
},
|
|
{
|
|
"id": 1,
|
|
"name": "echo"
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 2
|
|
}
|
|
"###);
|
|
let (documents, _code) = index.get_document(0, None).await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"id": 0,
|
|
"name": "kefir"
|
|
}
|
|
"###);
|
|
|
|
// if we try to retrieve the vectors with the `fields` parameter they
|
|
// still shouldn't be displayed
|
|
let (documents, _code) = index
|
|
.get_all_documents(GetAllDocumentsOptions {
|
|
fields: Some(vec!["name", "_vectors"]),
|
|
..Default::default()
|
|
})
|
|
.await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"name": "kefir"
|
|
},
|
|
{
|
|
"name": "echo"
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 2
|
|
}
|
|
"###);
|
|
let (documents, _code) =
|
|
index.get_document(0, Some(json!({"fields": ["name", "_vectors"]}))).await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"name": "kefir"
|
|
}
|
|
"###);
|
|
|
|
// If we specify the retrieve vectors boolean and nothing else we should get the vectors
|
|
let (documents, _code) = index
|
|
.get_all_documents(GetAllDocumentsOptions { retrieve_vectors: true, ..Default::default() })
|
|
.await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"id": 0,
|
|
"name": "kefir",
|
|
"_vectors": {
|
|
"manual": {
|
|
"embeddings": [
|
|
[
|
|
0.0,
|
|
0.0,
|
|
0.0
|
|
]
|
|
],
|
|
"regenerate": false
|
|
}
|
|
}
|
|
},
|
|
{
|
|
"id": 1,
|
|
"name": "echo",
|
|
"_vectors": {
|
|
"manual": {
|
|
"embeddings": [],
|
|
"regenerate": false
|
|
}
|
|
}
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 2
|
|
}
|
|
"###);
|
|
let (documents, _code) = index.get_document(0, Some(json!({"retrieveVectors": true}))).await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"id": 0,
|
|
"name": "kefir",
|
|
"_vectors": {
|
|
"manual": {
|
|
"embeddings": [
|
|
[
|
|
0.0,
|
|
0.0,
|
|
0.0
|
|
]
|
|
],
|
|
"regenerate": false
|
|
}
|
|
}
|
|
}
|
|
"###);
|
|
|
|
// If we specify the retrieve vectors boolean and exclude vectors form the `fields` we should still get the vectors
|
|
let (documents, _code) = index
|
|
.get_all_documents(GetAllDocumentsOptions {
|
|
retrieve_vectors: true,
|
|
fields: Some(vec!["name"]),
|
|
..Default::default()
|
|
})
|
|
.await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"results": [
|
|
{
|
|
"name": "kefir",
|
|
"_vectors": {
|
|
"manual": {
|
|
"embeddings": [
|
|
[
|
|
0.0,
|
|
0.0,
|
|
0.0
|
|
]
|
|
],
|
|
"regenerate": false
|
|
}
|
|
}
|
|
},
|
|
{
|
|
"name": "echo",
|
|
"_vectors": {
|
|
"manual": {
|
|
"embeddings": [],
|
|
"regenerate": false
|
|
}
|
|
}
|
|
}
|
|
],
|
|
"offset": 0,
|
|
"limit": 20,
|
|
"total": 2
|
|
}
|
|
"###);
|
|
let (documents, _code) =
|
|
index.get_document(0, Some(json!({"retrieveVectors": true, "fields": ["name"]}))).await;
|
|
snapshot!(json_string!(documents), @r###"
|
|
{
|
|
"name": "kefir",
|
|
"_vectors": {
|
|
"manual": {
|
|
"embeddings": [
|
|
[
|
|
0.0,
|
|
0.0,
|
|
0.0
|
|
]
|
|
],
|
|
"regenerate": false
|
|
}
|
|
}
|
|
}
|
|
"###);
|
|
}
|