use actix_web::error::PayloadError; use actix_web::http::header::CONTENT_TYPE; use actix_web::web::Bytes; use actix_web::HttpMessage; use actix_web::{web, HttpRequest, HttpResponse}; use bstr::ByteSlice; use futures::{Stream, StreamExt}; use log::debug; use meilisearch_lib::index_controller::{DocumentAdditionFormat, Update}; use meilisearch_lib::milli::update::IndexDocumentsMethod; use meilisearch_lib::MeiliSearch; use meilisearch_types::error::ResponseError; use mime::Mime; use once_cell::sync::Lazy; use serde::Deserialize; use serde_cs::vec::CS; use serde_json::Value; use tokio::sync::mpsc; use crate::analytics::Analytics; use crate::error::MeilisearchHttpError; use crate::extractors::authentication::{policies::*, GuardedData}; use crate::extractors::payload::Payload; use crate::extractors::sequential_extractor::SeqHandler; use crate::routes::{fold_star_or, PaginationView, StarOr}; use crate::task::SummarizedTaskView; static ACCEPTED_CONTENT_TYPE: Lazy> = Lazy::new(|| { vec![ "application/json".to_string(), "application/x-ndjson".to_string(), "text/csv".to_string(), ] }); /// This is required because Payload is not Sync nor Send fn payload_to_stream(mut payload: Payload) -> impl Stream> { let (snd, recv) = mpsc::channel(1); tokio::task::spawn_local(async move { while let Some(data) = payload.next().await { let _ = snd.send(data).await; } }); tokio_stream::wrappers::ReceiverStream::new(recv) } /// Extracts the mime type from the content type and return /// a meilisearch error if anyhthing bad happen. fn extract_mime_type(req: &HttpRequest) -> Result, MeilisearchHttpError> { match req.mime_type() { Ok(Some(mime)) => Ok(Some(mime)), Ok(None) => Ok(None), Err(_) => match req.headers().get(CONTENT_TYPE) { Some(content_type) => Err(MeilisearchHttpError::InvalidContentType( content_type.as_bytes().as_bstr().to_string(), ACCEPTED_CONTENT_TYPE.clone(), )), None => Err(MeilisearchHttpError::MissingContentType( ACCEPTED_CONTENT_TYPE.clone(), )), }, } } #[derive(Deserialize)] pub struct DocumentParam { index_uid: String, document_id: String, } pub fn configure(cfg: &mut web::ServiceConfig) { cfg.service( web::resource("") .route(web::get().to(SeqHandler(get_all_documents))) .route(web::post().to(SeqHandler(add_documents))) .route(web::put().to(SeqHandler(update_documents))) .route(web::delete().to(SeqHandler(clear_all_documents))), ) // this route needs to be before the /documents/{document_id} to match properly .service(web::resource("/delete-batch").route(web::post().to(SeqHandler(delete_documents)))) .service( web::resource("/{document_id}") .route(web::get().to(SeqHandler(get_document))) .route(web::delete().to(SeqHandler(delete_document))), ); } #[derive(Deserialize, Debug)] #[serde(rename_all = "camelCase", deny_unknown_fields)] pub struct GetDocument { fields: Option>>, } pub async fn get_document( meilisearch: GuardedData, MeiliSearch>, path: web::Path, params: web::Query, ) -> Result { let index = path.index_uid.clone(); let id = path.document_id.clone(); let GetDocument { fields } = params.into_inner(); let attributes_to_retrieve = fields.and_then(fold_star_or); let document = meilisearch .document(index, id, attributes_to_retrieve) .await?; debug!("returns: {:?}", document); Ok(HttpResponse::Ok().json(document)) } pub async fn delete_document( meilisearch: GuardedData, MeiliSearch>, path: web::Path, ) -> Result { let DocumentParam { document_id, index_uid, } = path.into_inner(); let update = Update::DeleteDocuments(vec![document_id]); let task: SummarizedTaskView = meilisearch.register_update(index_uid, update).await?.into(); debug!("returns: {:?}", task); Ok(HttpResponse::Accepted().json(task)) } #[derive(Deserialize, Debug)] #[serde(rename_all = "camelCase", deny_unknown_fields)] pub struct BrowseQuery { #[serde(default)] offset: usize, #[serde(default = "crate::routes::PAGINATION_DEFAULT_LIMIT")] limit: usize, fields: Option>>, } pub async fn get_all_documents( meilisearch: GuardedData, MeiliSearch>, path: web::Path, params: web::Query, ) -> Result { debug!("called with params: {:?}", params); let BrowseQuery { limit, offset, fields, } = params.into_inner(); let attributes_to_retrieve = fields.and_then(fold_star_or); let (total, documents) = meilisearch .documents(path.into_inner(), offset, limit, attributes_to_retrieve) .await?; let ret = PaginationView::new(offset, limit, total as usize, documents); debug!("returns: {:?}", ret); Ok(HttpResponse::Ok().json(ret)) } #[derive(Deserialize, Debug)] #[serde(rename_all = "camelCase", deny_unknown_fields)] pub struct UpdateDocumentsQuery { pub primary_key: Option, } pub async fn add_documents( meilisearch: GuardedData, MeiliSearch>, path: web::Path, params: web::Query, body: Payload, req: HttpRequest, analytics: web::Data, ) -> Result { debug!("called with params: {:?}", params); let params = params.into_inner(); let index_uid = path.into_inner(); analytics.add_documents( ¶ms, meilisearch.get_index(index_uid.clone()).await.is_err(), &req, ); let allow_index_creation = meilisearch.filters().allow_index_creation; let task = document_addition( extract_mime_type(&req)?, meilisearch, index_uid, params.primary_key, body, IndexDocumentsMethod::ReplaceDocuments, allow_index_creation, ) .await?; Ok(HttpResponse::Accepted().json(task)) } pub async fn update_documents( meilisearch: GuardedData, MeiliSearch>, path: web::Path, params: web::Query, body: Payload, req: HttpRequest, analytics: web::Data, ) -> Result { debug!("called with params: {:?}", params); let index_uid = path.into_inner(); analytics.update_documents( ¶ms, meilisearch.get_index(index_uid.clone()).await.is_err(), &req, ); let allow_index_creation = meilisearch.filters().allow_index_creation; let task = document_addition( extract_mime_type(&req)?, meilisearch, index_uid, params.into_inner().primary_key, body, IndexDocumentsMethod::UpdateDocuments, allow_index_creation, ) .await?; Ok(HttpResponse::Accepted().json(task)) } async fn document_addition( mime_type: Option, meilisearch: GuardedData, MeiliSearch>, index_uid: String, primary_key: Option, body: Payload, method: IndexDocumentsMethod, allow_index_creation: bool, ) -> Result { let format = match mime_type .as_ref() .map(|m| (m.type_().as_str(), m.subtype().as_str())) { Some(("application", "json")) => DocumentAdditionFormat::Json, Some(("application", "x-ndjson")) => DocumentAdditionFormat::Ndjson, Some(("text", "csv")) => DocumentAdditionFormat::Csv, Some((type_, subtype)) => { return Err(MeilisearchHttpError::InvalidContentType( format!("{}/{}", type_, subtype), ACCEPTED_CONTENT_TYPE.clone(), ) .into()) } None => { return Err( MeilisearchHttpError::MissingContentType(ACCEPTED_CONTENT_TYPE.clone()).into(), ) } }; let update = Update::DocumentAddition { payload: Box::new(payload_to_stream(body)), primary_key, method, format, allow_index_creation, }; let task = meilisearch.register_update(index_uid, update).await?.into(); debug!("returns: {:?}", task); Ok(task) } pub async fn delete_documents( meilisearch: GuardedData, MeiliSearch>, path: web::Path, body: web::Json>, ) -> Result { debug!("called with params: {:?}", body); let ids = body .iter() .map(|v| { v.as_str() .map(String::from) .unwrap_or_else(|| v.to_string()) }) .collect(); let update = Update::DeleteDocuments(ids); let task: SummarizedTaskView = meilisearch .register_update(path.into_inner(), update) .await? .into(); debug!("returns: {:?}", task); Ok(HttpResponse::Accepted().json(task)) } pub async fn clear_all_documents( meilisearch: GuardedData, MeiliSearch>, path: web::Path, ) -> Result { let update = Update::ClearDocuments; let task: SummarizedTaskView = meilisearch .register_update(path.into_inner(), update) .await? .into(); debug!("returns: {:?}", task); Ok(HttpResponse::Accepted().json(task)) }