mirror of
https://github.com/meilisearch/MeiliSearch
synced 2024-11-25 22:34:28 +01:00
Apply PR review comments
This commit is contained in:
parent
a690ace36e
commit
4dfae44478
156
config.toml
156
config.toml
@ -1,50 +1,55 @@
|
|||||||
# This file shows the default configuration of Meilisearch.
|
# This file shows the default configuration of Meilisearch.
|
||||||
# All variables are defined here https://docs.meilisearch.com/learn/configuration/instance_options.html#environment-variables
|
# All variables are defined here https://docs.meilisearch.com/learn/configuration/instance_options.html#environment-variables
|
||||||
|
|
||||||
|
# db_path = "./data.ms"
|
||||||
|
# The destination where the database must be created.
|
||||||
|
|
||||||
|
# env = "development" # Possible values: [development, production]
|
||||||
|
# This environment variable must be set to `production` if you are running in production.
|
||||||
|
# More logs wiil be displayed if the server is running in development mode. Setting the master
|
||||||
|
# key is optional; hence no security on the updates routes. This
|
||||||
|
# is useful to debug when integrating the engine with another service.
|
||||||
|
|
||||||
|
# http_addr = "127.0.0.1:7700"
|
||||||
|
# The address on which the HTTP server will listen.
|
||||||
|
|
||||||
|
# master-key = "MASTER_KEY"
|
||||||
|
# Sets the instance's master key, automatically protecting all routes except GET /health.
|
||||||
|
|
||||||
|
# no_analytics = false
|
||||||
|
# Do not send analytics to Meilisearch.
|
||||||
|
|
||||||
|
# disable-auto-batching = false
|
||||||
|
# The engine will disable task auto-batching, and will sequencialy compute each task one by one.
|
||||||
|
|
||||||
|
|
||||||
### DUMP
|
### DUMP
|
||||||
|
|
||||||
# Folder where dumps are created when the dump route is called
|
|
||||||
# dumps-dir = "dumps/"
|
# dumps-dir = "dumps/"
|
||||||
|
# Folder where dumps are created when the dump route is called.
|
||||||
|
|
||||||
# Ignore the dump if a database already exists, and load that database instead
|
|
||||||
# ignore-dump-if-db-exists = false
|
|
||||||
|
|
||||||
# If the dump doesn't exists, load or create the database specified by `db-path` instead
|
|
||||||
# ignore-missing-dump = false
|
|
||||||
|
|
||||||
# Import a dump from the specified path, must be a `.dump` file
|
|
||||||
# import-dump = "./path/to/my/file.dump"
|
# import-dump = "./path/to/my/file.dump"
|
||||||
|
# Import a dump from the specified path, must be a `.dump` file.
|
||||||
|
|
||||||
|
# ignore-missing-dump = false
|
||||||
|
# If the dump doesn't exist, load or create the database specified by `db-path` instead.
|
||||||
|
|
||||||
|
# ignore-dump-if-db-exists = false
|
||||||
|
# Ignore the dump if a database already exists, and load that database instead.
|
||||||
|
|
||||||
|
###
|
||||||
|
|
||||||
|
|
||||||
### SNAPSHOT
|
# log-level = "INFO" # Possible values: [ERROR, WARN, INFO, DEBUG, TRACE]
|
||||||
|
# Set the log level.
|
||||||
# The engine will ignore a missing snapshot and not return an error in such case
|
|
||||||
# ignore-missing-snapshot = false
|
|
||||||
|
|
||||||
# The engine will skip snapshot importation and not return an error in such case
|
|
||||||
# ignore-snapshot-if-db-exists = false
|
|
||||||
|
|
||||||
# Defines the path of the snapshot file to import. This option will, by default, stop the
|
|
||||||
# process if a database already exist or if no snapshot exists at the given path. If this
|
|
||||||
# option is not specified no snapshot is imported
|
|
||||||
# import-snapshot = false
|
|
||||||
|
|
||||||
# Activate snapshot scheduling
|
|
||||||
# schedule-snapshot = false
|
|
||||||
|
|
||||||
# Defines the directory path where meilisearch will create snapshot each snapshot_time_gap
|
|
||||||
# snapshot-dir = "snapshots/"
|
|
||||||
|
|
||||||
# Defines time interval, in seconds, between each snapshot creation
|
|
||||||
# snapshot-interval-sec = 86400
|
|
||||||
|
|
||||||
|
|
||||||
### INDEX
|
### INDEX
|
||||||
|
|
||||||
# The maximum size, in bytes, of the main lmdb database directory
|
|
||||||
# max-index-size = "100 GiB"
|
# max-index-size = "100 GiB"
|
||||||
|
# The maximum size, in bytes, of the main LMDB database directory.
|
||||||
|
|
||||||
|
# max-indexing-memory = "2 GiB"
|
||||||
# The maximum amount of memory the indexer will use. It defaults to 2/3 of the available
|
# The maximum amount of memory the indexer will use. It defaults to 2/3 of the available
|
||||||
# memory. It is recommended to use something like 80%-90% of the available memory, no
|
# memory. It is recommended to use something like 80%-90% of the available memory, no
|
||||||
# more.
|
# more.
|
||||||
@ -54,68 +59,73 @@
|
|||||||
# is recommended to specify the amount of memory to use.
|
# is recommended to specify the amount of memory to use.
|
||||||
#
|
#
|
||||||
# /!\ The default value is system dependant /!\
|
# /!\ The default value is system dependant /!\
|
||||||
# max-indexing-memory = "2 GiB"
|
|
||||||
|
|
||||||
|
# max-indexing-threads = 4
|
||||||
# The maximum number of threads the indexer will use. If the number set is higher than the
|
# The maximum number of threads the indexer will use. If the number set is higher than the
|
||||||
# real number of cores available in the machine, it will use the maximum number of
|
# real number of cores available in the machine, it will use the maximum number of
|
||||||
# available cores.
|
# available cores.
|
||||||
#
|
#
|
||||||
# It defaults to half of the available threads.
|
# It defaults to half of the available threads.
|
||||||
# max-indexing-threads = 4
|
|
||||||
|
###
|
||||||
|
|
||||||
|
|
||||||
|
# max-task-db-size = "100 GiB"
|
||||||
|
# The maximum size, in bytes, of the update LMDB database directory.
|
||||||
|
|
||||||
|
# http-payload-size-limit = 100000000
|
||||||
|
# The maximum size, in bytes, of accepted JSON payloads.
|
||||||
|
|
||||||
|
|
||||||
|
### SNAPSHOT
|
||||||
|
|
||||||
|
# schedule-snapshot = false
|
||||||
|
# Activate snapshot scheduling.
|
||||||
|
|
||||||
|
# snapshot-dir = "snapshots/"
|
||||||
|
# Defines the directory path where Meilisearch will create a snapshot each snapshot-interval-sec.
|
||||||
|
|
||||||
|
# snapshot-interval-sec = 86400
|
||||||
|
# Defines time interval, in seconds, between each snapshot creation.
|
||||||
|
|
||||||
|
# import-snapshot = false
|
||||||
|
# Defines the path of the snapshot file to import. This option will, by default, stop the
|
||||||
|
# process if a database already exist, or if no snapshot exists at the given path. If this
|
||||||
|
# option is not specified, no snapshot is imported.
|
||||||
|
|
||||||
|
# ignore-missing-snapshot = false
|
||||||
|
# The engine will ignore a missing snapshot and not return an error in such a case.
|
||||||
|
|
||||||
|
# ignore-snapshot-if-db-exists = false
|
||||||
|
# The engine will skip snapshot importation and not return an error in such a case.
|
||||||
|
|
||||||
|
###
|
||||||
|
|
||||||
|
|
||||||
### SSL
|
### SSL
|
||||||
|
|
||||||
# Enable client authentication, and accept certificates signed by those roots provided in CERTFILE
|
|
||||||
# ssl-auth-path = "./path/to/root"
|
# ssl-auth-path = "./path/to/root"
|
||||||
|
# Enable client authentication, and accept certificates signed by those roots provided in CERTFILE.
|
||||||
|
|
||||||
|
# ssl-cert-path = "./path/to/CERTFILE"
|
||||||
# Read server certificates from CERTFILE. This should contain PEM-format certificates in
|
# Read server certificates from CERTFILE. This should contain PEM-format certificates in
|
||||||
# the right order (the first certificate should certify KEYFILE, the last should be a root
|
# the right order (the first certificate should certify KEYFILE, the last should be a root
|
||||||
# CA)
|
# CA).
|
||||||
# ssl-cert-path = "./path/to/CERTFILE"
|
|
||||||
|
|
||||||
# Read private key from KEYFILE. This should be a RSA private key or PKCS8-encoded
|
|
||||||
# private key, in PEM format
|
|
||||||
# ssl-key-path = "./path/to/private-key"
|
# ssl-key-path = "./path/to/private-key"
|
||||||
|
# Read the private key from KEYFILE. This should be an RSA private key or PKCS8-encoded
|
||||||
# Read DER-encoded OCSP response from OCSPFILE and staple to certificate. Optional
|
# private key, in PEM format.
|
||||||
|
|
||||||
# ssl-ocsp-path = "./path/to/OCSPFILE"
|
# ssl-ocsp-path = "./path/to/OCSPFILE"
|
||||||
|
# Read DER-encoded OCSP response from OCSPFILE and staple to certificate. Optional.
|
||||||
|
|
||||||
# Send a fatal alert if the client does not complete client authentication
|
|
||||||
# ssl-require-auth = false
|
# ssl-require-auth = false
|
||||||
|
# Send a fatal alert if the client does not complete client authentication.
|
||||||
# SSL support session resumption
|
|
||||||
# ssl-resumption = false
|
# ssl-resumption = false
|
||||||
|
# SSL support session resumption.
|
||||||
# SSL support tickets
|
|
||||||
# ssl-tickets = false
|
# ssl-tickets = false
|
||||||
|
# SSL support tickets.
|
||||||
|
|
||||||
|
###
|
||||||
### MISC
|
|
||||||
|
|
||||||
# This environment variable must be set to `production` if you are running in production.
|
|
||||||
# If the server is running in development mode more logs will be displayed, and the master
|
|
||||||
# key can be avoided which implies that there is no security on the updates routes. This
|
|
||||||
# is useful to debug when integrating the engine with another service
|
|
||||||
# env = "development" # possible values: [development, production]
|
|
||||||
|
|
||||||
# The address on which the http server will listen
|
|
||||||
# http-addr = "127.0.0.1:7700"
|
|
||||||
|
|
||||||
# The maximum size, in bytes, of accepted JSON payloads
|
|
||||||
# http-payload-size-limit = 100000000
|
|
||||||
|
|
||||||
# The destination where the database must be created
|
|
||||||
# db-path = "./data.ms"
|
|
||||||
|
|
||||||
# The engine will disable task auto-batching, and will sequencialy compute each task one by one
|
|
||||||
# disable-auto-batching = false
|
|
||||||
|
|
||||||
# Set the log level
|
|
||||||
# log-level = "info"
|
|
||||||
|
|
||||||
# The master key allowing you to do everything on the server
|
|
||||||
# master-key = "YOUR MASTER KEY"
|
|
||||||
|
|
||||||
# The maximum size, in bytes, of the update lmdb database directory
|
|
||||||
# max-task-db-size = "100 GiB"
|
|
||||||
|
@ -1,4 +1,5 @@
|
|||||||
use std::env;
|
use std::env;
|
||||||
|
use std::path::PathBuf;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
|
|
||||||
use actix_web::http::KeepAlive;
|
use actix_web::http::KeepAlive;
|
||||||
@ -28,7 +29,7 @@ fn setup(opt: &Opt) -> anyhow::Result<()> {
|
|||||||
|
|
||||||
#[actix_web::main]
|
#[actix_web::main]
|
||||||
async fn main() -> anyhow::Result<()> {
|
async fn main() -> anyhow::Result<()> {
|
||||||
let opt = Opt::build();
|
let (opt, config_read_from) = Opt::try_build()?;
|
||||||
|
|
||||||
setup(&opt)?;
|
setup(&opt)?;
|
||||||
|
|
||||||
@ -57,7 +58,7 @@ async fn main() -> anyhow::Result<()> {
|
|||||||
#[cfg(any(debug_assertions, not(feature = "analytics")))]
|
#[cfg(any(debug_assertions, not(feature = "analytics")))]
|
||||||
let (analytics, user) = analytics::MockAnalytics::new(&opt);
|
let (analytics, user) = analytics::MockAnalytics::new(&opt);
|
||||||
|
|
||||||
print_launch_resume(&opt, &user);
|
print_launch_resume(&opt, &user, config_read_from);
|
||||||
|
|
||||||
run_http(meilisearch, auth_controller, opt, analytics).await?;
|
run_http(meilisearch, auth_controller, opt, analytics).await?;
|
||||||
|
|
||||||
@ -96,7 +97,7 @@ async fn run_http(
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn print_launch_resume(opt: &Opt, user: &str) {
|
pub fn print_launch_resume(opt: &Opt, user: &str, config_read_from: Option<PathBuf>) {
|
||||||
let commit_sha = option_env!("VERGEN_GIT_SHA").unwrap_or("unknown");
|
let commit_sha = option_env!("VERGEN_GIT_SHA").unwrap_or("unknown");
|
||||||
let commit_date = option_env!("VERGEN_GIT_COMMIT_TIMESTAMP").unwrap_or("unknown");
|
let commit_date = option_env!("VERGEN_GIT_COMMIT_TIMESTAMP").unwrap_or("unknown");
|
||||||
let protocol = if opt.ssl_cert_path.is_some() && opt.ssl_key_path.is_some() {
|
let protocol = if opt.ssl_cert_path.is_some() && opt.ssl_key_path.is_some() {
|
||||||
@ -117,6 +118,12 @@ pub fn print_launch_resume(opt: &Opt, user: &str) {
|
|||||||
|
|
||||||
eprintln!("{}", ascii_name);
|
eprintln!("{}", ascii_name);
|
||||||
|
|
||||||
|
eprintln!(
|
||||||
|
"Config file path:\t{}",
|
||||||
|
config_read_from
|
||||||
|
.map(|config_file_path| config_file_path.display().to_string())
|
||||||
|
.unwrap_or_else(|| "none".to_string())
|
||||||
|
);
|
||||||
eprintln!("Database path:\t\t{:?}", opt.db_path);
|
eprintln!("Database path:\t\t{:?}", opt.db_path);
|
||||||
eprintln!("Server listening on:\t\"{}://{}\"", protocol, opt.http_addr);
|
eprintln!("Server listening on:\t\"{}://{}\"", protocol, opt.http_addr);
|
||||||
eprintln!("Environment:\t\t{:?}", opt.env);
|
eprintln!("Environment:\t\t{:?}", opt.env);
|
||||||
|
@ -64,6 +64,7 @@ const DEFAULT_LOG_LEVEL: &str = "info";
|
|||||||
|
|
||||||
#[derive(Debug, Clone, Parser, Serialize, Deserialize)]
|
#[derive(Debug, Clone, Parser, Serialize, Deserialize)]
|
||||||
#[clap(version)]
|
#[clap(version)]
|
||||||
|
#[serde(rename_all = "snake_case", deny_unknown_fields)]
|
||||||
pub struct Opt {
|
pub struct Opt {
|
||||||
/// The destination where the database must be created.
|
/// The destination where the database must be created.
|
||||||
#[clap(long, env = MEILI_DB_PATH, default_value_os_t = default_db_path())]
|
#[clap(long, env = MEILI_DB_PATH, default_value_os_t = default_db_path())]
|
||||||
@ -75,15 +76,15 @@ pub struct Opt {
|
|||||||
#[serde(default = "default_http_addr")]
|
#[serde(default = "default_http_addr")]
|
||||||
pub http_addr: String,
|
pub http_addr: String,
|
||||||
|
|
||||||
/// The master key allowing you to do everything on the server.
|
/// Sets the instance's master key, automatically protecting all routes except GET /health
|
||||||
#[serde(skip_serializing)]
|
#[serde(skip_serializing)]
|
||||||
#[clap(long, env = MEILI_MASTER_KEY)]
|
#[clap(long, env = MEILI_MASTER_KEY)]
|
||||||
pub master_key: Option<String>,
|
pub master_key: Option<String>,
|
||||||
|
|
||||||
/// This environment variable must be set to `production` if you are running in production.
|
/// This environment variable must be set to `production` if you are running in production.
|
||||||
/// If the server is running in development mode more logs will be displayed,
|
/// More logs wiil be displayed if the server is running in development mode. Setting the master
|
||||||
/// and the master key can be avoided which implies that there is no security on the updates routes.
|
/// key is optional; hence no security on the updates routes. This
|
||||||
/// This is useful to debug when integrating the engine with another service.
|
/// is useful to debug when integrating the engine with another service
|
||||||
#[clap(long, env = MEILI_ENV, default_value_t = default_env(), possible_values = &POSSIBLE_ENV)]
|
#[clap(long, env = MEILI_ENV, default_value_t = default_env(), possible_values = &POSSIBLE_ENV)]
|
||||||
#[serde(default = "default_env")]
|
#[serde(default = "default_env")]
|
||||||
pub env: String,
|
pub env: String,
|
||||||
@ -94,12 +95,12 @@ pub struct Opt {
|
|||||||
#[clap(long, env = MEILI_NO_ANALYTICS)]
|
#[clap(long, env = MEILI_NO_ANALYTICS)]
|
||||||
pub no_analytics: bool,
|
pub no_analytics: bool,
|
||||||
|
|
||||||
/// The maximum size, in bytes, of the main lmdb database directory
|
/// The maximum size, in bytes, of the main LMDB database directory
|
||||||
#[clap(long, env = MEILI_MAX_INDEX_SIZE, default_value_t = default_max_index_size())]
|
#[clap(long, env = MEILI_MAX_INDEX_SIZE, default_value_t = default_max_index_size())]
|
||||||
#[serde(default = "default_max_index_size")]
|
#[serde(default = "default_max_index_size")]
|
||||||
pub max_index_size: Byte,
|
pub max_index_size: Byte,
|
||||||
|
|
||||||
/// The maximum size, in bytes, of the update lmdb database directory
|
/// The maximum size, in bytes, of the update LMDB database directory
|
||||||
#[clap(long, env = MEILI_MAX_TASK_DB_SIZE, default_value_t = default_max_task_db_size())]
|
#[clap(long, env = MEILI_MAX_TASK_DB_SIZE, default_value_t = default_max_task_db_size())]
|
||||||
#[serde(default = "default_max_task_db_size")]
|
#[serde(default = "default_max_task_db_size")]
|
||||||
pub max_task_db_size: Byte,
|
pub max_task_db_size: Byte,
|
||||||
@ -117,7 +118,7 @@ pub struct Opt {
|
|||||||
#[clap(long, env = MEILI_SSL_CERT_PATH, parse(from_os_str))]
|
#[clap(long, env = MEILI_SSL_CERT_PATH, parse(from_os_str))]
|
||||||
pub ssl_cert_path: Option<PathBuf>,
|
pub ssl_cert_path: Option<PathBuf>,
|
||||||
|
|
||||||
/// Read private key from KEYFILE. This should be a RSA
|
/// Read the private key from KEYFILE. This should be an RSA
|
||||||
/// private key or PKCS8-encoded private key, in PEM format.
|
/// private key or PKCS8-encoded private key, in PEM format.
|
||||||
#[serde(skip_serializing)]
|
#[serde(skip_serializing)]
|
||||||
#[clap(long, env = MEILI_SSL_KEY_PATH, parse(from_os_str))]
|
#[clap(long, env = MEILI_SSL_KEY_PATH, parse(from_os_str))]
|
||||||
@ -151,12 +152,12 @@ pub struct Opt {
|
|||||||
pub ssl_tickets: bool,
|
pub ssl_tickets: bool,
|
||||||
|
|
||||||
/// Defines the path of the snapshot file to import.
|
/// Defines the path of the snapshot file to import.
|
||||||
/// This option will, by default, stop the process if a database already exist or if no snapshot exists at
|
/// This option will, by default, stop the process if a database already exists or if no snapshot exists at
|
||||||
/// the given path. If this option is not specified no snapshot is imported.
|
/// the given path. If this option is not specified, no snapshot is imported.
|
||||||
#[clap(long, env = MEILI_IMPORT_SNAPSHOT)]
|
#[clap(long, env = MEILI_IMPORT_SNAPSHOT)]
|
||||||
pub import_snapshot: Option<PathBuf>,
|
pub import_snapshot: Option<PathBuf>,
|
||||||
|
|
||||||
/// The engine will ignore a missing snapshot and not return an error in such case.
|
/// The engine will ignore a missing snapshot and not return an error in such a case.
|
||||||
#[clap(
|
#[clap(
|
||||||
long,
|
long,
|
||||||
env = MEILI_IGNORE_MISSING_SNAPSHOT,
|
env = MEILI_IGNORE_MISSING_SNAPSHOT,
|
||||||
@ -174,7 +175,7 @@ pub struct Opt {
|
|||||||
#[serde(default)]
|
#[serde(default)]
|
||||||
pub ignore_snapshot_if_db_exists: bool,
|
pub ignore_snapshot_if_db_exists: bool,
|
||||||
|
|
||||||
/// Defines the directory path where meilisearch will create snapshot each snapshot_time_gap.
|
/// Defines the directory path where Meilisearch will create a snapshot each snapshot-interval-sec.
|
||||||
#[clap(long, env = MEILI_SNAPSHOT_DIR, default_value_os_t = default_snapshot_dir())]
|
#[clap(long, env = MEILI_SNAPSHOT_DIR, default_value_os_t = default_snapshot_dir())]
|
||||||
#[serde(default = "default_snapshot_dir")]
|
#[serde(default = "default_snapshot_dir")]
|
||||||
pub snapshot_dir: PathBuf,
|
pub snapshot_dir: PathBuf,
|
||||||
@ -194,7 +195,7 @@ pub struct Opt {
|
|||||||
#[clap(long, env = MEILI_IMPORT_DUMP, conflicts_with = "import-snapshot")]
|
#[clap(long, env = MEILI_IMPORT_DUMP, conflicts_with = "import-snapshot")]
|
||||||
pub import_dump: Option<PathBuf>,
|
pub import_dump: Option<PathBuf>,
|
||||||
|
|
||||||
/// If the dump doesn't exists, load or create the database specified by `db-path` instead.
|
/// If the dump doesn't exist, load or create the database specified by `db-path` instead.
|
||||||
#[clap(long, env = MEILI_IGNORE_MISSING_DUMP, requires = "import-dump")]
|
#[clap(long, env = MEILI_IGNORE_MISSING_DUMP, requires = "import-dump")]
|
||||||
#[serde(default)]
|
#[serde(default)]
|
||||||
pub ignore_missing_dump: bool,
|
pub ignore_missing_dump: bool,
|
||||||
@ -209,7 +210,7 @@ pub struct Opt {
|
|||||||
#[serde(default = "default_dumps_dir")]
|
#[serde(default = "default_dumps_dir")]
|
||||||
pub dumps_dir: PathBuf,
|
pub dumps_dir: PathBuf,
|
||||||
|
|
||||||
/// Set the log level
|
/// Set the log level. # Possible values: [ERROR, WARN, INFO, DEBUG, TRACE]
|
||||||
#[clap(long, env = MEILI_LOG_LEVEL, default_value_t = default_log_level())]
|
#[clap(long, env = MEILI_LOG_LEVEL, default_value_t = default_log_level())]
|
||||||
#[serde(default = "default_log_level")]
|
#[serde(default = "default_log_level")]
|
||||||
pub log_level: String,
|
pub log_level: String,
|
||||||
@ -243,78 +244,124 @@ impl Opt {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Build a new Opt from config file, env vars and cli args.
|
/// Build a new Opt from config file, env vars and cli args.
|
||||||
pub fn build() -> Self {
|
pub fn try_build() -> anyhow::Result<(Self, Option<PathBuf>)> {
|
||||||
// Parse the args to get the config_file_path.
|
// Parse the args to get the config_file_path.
|
||||||
let mut opts = Opt::parse();
|
let mut opts = Opt::parse();
|
||||||
if let Some(config_file_path) = opts.config_file_path.as_ref() {
|
let mut config_read_from = None;
|
||||||
eprintln!("loading config file : {:?}", config_file_path);
|
if let Some(config_file_path) = opts
|
||||||
match std::fs::read(config_file_path) {
|
.config_file_path
|
||||||
|
.clone()
|
||||||
|
.or_else(|| Some(PathBuf::from("./config.toml")))
|
||||||
|
{
|
||||||
|
match std::fs::read(&config_file_path) {
|
||||||
Ok(config) => {
|
Ok(config) => {
|
||||||
// If the arg is present, and the file successfully read, we deserialize it with `toml`.
|
// If the file is successfully read, we deserialize it with `toml`.
|
||||||
let opt_from_config =
|
match toml::from_slice::<Opt>(&config) {
|
||||||
toml::from_slice::<Opt>(&config).expect("can't read file");
|
Ok(opt_from_config) => {
|
||||||
// We inject the values from the toml in the corresponding env vars if needs be. Doing so, we respect the priority toml < env vars < cli args.
|
// We inject the values from the toml in the corresponding env vars if needs be. Doing so, we respect the priority toml < env vars < cli args.
|
||||||
opt_from_config.export_to_env();
|
opt_from_config.export_to_env();
|
||||||
// Once injected we parse the cli args once again to take the new env vars into scope.
|
// Once injected we parse the cli args once again to take the new env vars into scope.
|
||||||
opts = Opt::parse();
|
opts = Opt::parse();
|
||||||
|
config_read_from = Some(config_file_path);
|
||||||
|
}
|
||||||
|
// If we have an error deserializing the file defined by the user.
|
||||||
|
Err(err) if opts.config_file_path.is_some() => anyhow::bail!(err),
|
||||||
|
_ => (),
|
||||||
|
}
|
||||||
}
|
}
|
||||||
Err(err) => eprintln!("can't read {:?} : {}", config_file_path, err),
|
// If we have an error while reading the file defined by the user.
|
||||||
|
Err(err) if opts.config_file_path.is_some() => anyhow::bail!(err),
|
||||||
|
_ => (),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
opts
|
Ok((opts, config_read_from))
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Exports the opts values to their corresponding env vars if they are not set.
|
/// Exports the opts values to their corresponding env vars if they are not set.
|
||||||
fn export_to_env(self) {
|
fn export_to_env(self) {
|
||||||
export_to_env_if_not_present(MEILI_DB_PATH, self.db_path);
|
let Opt {
|
||||||
export_to_env_if_not_present(MEILI_HTTP_ADDR, self.http_addr);
|
db_path,
|
||||||
if let Some(master_key) = self.master_key {
|
http_addr,
|
||||||
|
master_key,
|
||||||
|
env,
|
||||||
|
max_index_size,
|
||||||
|
max_task_db_size,
|
||||||
|
http_payload_size_limit,
|
||||||
|
ssl_cert_path,
|
||||||
|
ssl_key_path,
|
||||||
|
ssl_auth_path,
|
||||||
|
ssl_ocsp_path,
|
||||||
|
ssl_require_auth,
|
||||||
|
ssl_resumption,
|
||||||
|
ssl_tickets,
|
||||||
|
snapshot_dir,
|
||||||
|
schedule_snapshot,
|
||||||
|
snapshot_interval_sec,
|
||||||
|
dumps_dir,
|
||||||
|
log_level,
|
||||||
|
indexer_options,
|
||||||
|
scheduler_options,
|
||||||
|
import_snapshot: _,
|
||||||
|
ignore_missing_snapshot: _,
|
||||||
|
ignore_snapshot_if_db_exists: _,
|
||||||
|
import_dump: _,
|
||||||
|
ignore_missing_dump: _,
|
||||||
|
ignore_dump_if_db_exists: _,
|
||||||
|
config_file_path: _,
|
||||||
|
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||||
|
no_analytics,
|
||||||
|
#[cfg(feature = "metrics")]
|
||||||
|
enable_metrics_route,
|
||||||
|
} = self;
|
||||||
|
export_to_env_if_not_present(MEILI_DB_PATH, db_path);
|
||||||
|
export_to_env_if_not_present(MEILI_HTTP_ADDR, http_addr);
|
||||||
|
if let Some(master_key) = master_key {
|
||||||
export_to_env_if_not_present(MEILI_MASTER_KEY, master_key);
|
export_to_env_if_not_present(MEILI_MASTER_KEY, master_key);
|
||||||
}
|
}
|
||||||
export_to_env_if_not_present(MEILI_ENV, self.env);
|
export_to_env_if_not_present(MEILI_ENV, env);
|
||||||
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||||
{
|
{
|
||||||
export_to_env_if_not_present(MEILI_NO_ANALYTICS, self.no_analytics.to_string());
|
export_to_env_if_not_present(MEILI_NO_ANALYTICS, no_analytics.to_string());
|
||||||
}
|
}
|
||||||
export_to_env_if_not_present(MEILI_MAX_INDEX_SIZE, self.max_index_size.to_string());
|
export_to_env_if_not_present(MEILI_MAX_INDEX_SIZE, max_index_size.to_string());
|
||||||
export_to_env_if_not_present(MEILI_MAX_TASK_DB_SIZE, self.max_task_db_size.to_string());
|
export_to_env_if_not_present(MEILI_MAX_TASK_DB_SIZE, max_task_db_size.to_string());
|
||||||
export_to_env_if_not_present(
|
export_to_env_if_not_present(
|
||||||
MEILI_HTTP_PAYLOAD_SIZE_LIMIT,
|
MEILI_HTTP_PAYLOAD_SIZE_LIMIT,
|
||||||
self.http_payload_size_limit.to_string(),
|
http_payload_size_limit.to_string(),
|
||||||
);
|
);
|
||||||
if let Some(ssl_cert_path) = self.ssl_cert_path {
|
if let Some(ssl_cert_path) = ssl_cert_path {
|
||||||
export_to_env_if_not_present(MEILI_SSL_CERT_PATH, ssl_cert_path);
|
export_to_env_if_not_present(MEILI_SSL_CERT_PATH, ssl_cert_path);
|
||||||
}
|
}
|
||||||
if let Some(ssl_key_path) = self.ssl_key_path {
|
if let Some(ssl_key_path) = ssl_key_path {
|
||||||
export_to_env_if_not_present(MEILI_SSL_KEY_PATH, ssl_key_path);
|
export_to_env_if_not_present(MEILI_SSL_KEY_PATH, ssl_key_path);
|
||||||
}
|
}
|
||||||
if let Some(ssl_auth_path) = self.ssl_auth_path {
|
if let Some(ssl_auth_path) = ssl_auth_path {
|
||||||
export_to_env_if_not_present(MEILI_SSL_AUTH_PATH, ssl_auth_path);
|
export_to_env_if_not_present(MEILI_SSL_AUTH_PATH, ssl_auth_path);
|
||||||
}
|
}
|
||||||
if let Some(ssl_ocsp_path) = self.ssl_ocsp_path {
|
if let Some(ssl_ocsp_path) = ssl_ocsp_path {
|
||||||
export_to_env_if_not_present(MEILI_SSL_OCSP_PATH, ssl_ocsp_path);
|
export_to_env_if_not_present(MEILI_SSL_OCSP_PATH, ssl_ocsp_path);
|
||||||
}
|
}
|
||||||
export_to_env_if_not_present(MEILI_SSL_REQUIRE_AUTH, self.ssl_require_auth.to_string());
|
export_to_env_if_not_present(MEILI_SSL_REQUIRE_AUTH, ssl_require_auth.to_string());
|
||||||
export_to_env_if_not_present(MEILI_SSL_RESUMPTION, self.ssl_resumption.to_string());
|
export_to_env_if_not_present(MEILI_SSL_RESUMPTION, ssl_resumption.to_string());
|
||||||
export_to_env_if_not_present(MEILI_SSL_TICKETS, self.ssl_tickets.to_string());
|
export_to_env_if_not_present(MEILI_SSL_TICKETS, ssl_tickets.to_string());
|
||||||
export_to_env_if_not_present(MEILI_SNAPSHOT_DIR, self.snapshot_dir);
|
export_to_env_if_not_present(MEILI_SNAPSHOT_DIR, snapshot_dir);
|
||||||
export_to_env_if_not_present(MEILI_SCHEDULE_SNAPSHOT, self.schedule_snapshot.to_string());
|
export_to_env_if_not_present(MEILI_SCHEDULE_SNAPSHOT, schedule_snapshot.to_string());
|
||||||
export_to_env_if_not_present(
|
export_to_env_if_not_present(
|
||||||
MEILI_SNAPSHOT_INTERVAL_SEC,
|
MEILI_SNAPSHOT_INTERVAL_SEC,
|
||||||
self.snapshot_interval_sec.to_string(),
|
snapshot_interval_sec.to_string(),
|
||||||
);
|
);
|
||||||
export_to_env_if_not_present(MEILI_DUMPS_DIR, self.dumps_dir);
|
export_to_env_if_not_present(MEILI_DUMPS_DIR, dumps_dir);
|
||||||
export_to_env_if_not_present(MEILI_LOG_LEVEL, self.log_level);
|
export_to_env_if_not_present(MEILI_LOG_LEVEL, log_level);
|
||||||
#[cfg(feature = "metrics")]
|
#[cfg(feature = "metrics")]
|
||||||
{
|
{
|
||||||
export_to_env_if_not_present(
|
export_to_env_if_not_present(
|
||||||
MEILI_ENABLE_METRICS_ROUTE,
|
MEILI_ENABLE_METRICS_ROUTE,
|
||||||
self.enable_metrics_route.to_string(),
|
enable_metrics_route.to_string(),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
self.indexer_options.export_to_env();
|
indexer_options.export_to_env();
|
||||||
self.scheduler_options.export_to_env();
|
scheduler_options.export_to_env();
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn get_ssl_config(&self) -> anyhow::Result<Option<rustls::ServerConfig>> {
|
pub fn get_ssl_config(&self) -> anyhow::Result<Option<rustls::ServerConfig>> {
|
||||||
|
@ -12,10 +12,10 @@ use sysinfo::{RefreshKind, System, SystemExt};
|
|||||||
const MEILI_MAX_INDEXING_MEMORY: &str = "MEILI_MAX_INDEXING_MEMORY";
|
const MEILI_MAX_INDEXING_MEMORY: &str = "MEILI_MAX_INDEXING_MEMORY";
|
||||||
const MEILI_MAX_INDEXING_THREADS: &str = "MEILI_MAX_INDEXING_THREADS";
|
const MEILI_MAX_INDEXING_THREADS: &str = "MEILI_MAX_INDEXING_THREADS";
|
||||||
const DISABLE_AUTO_BATCHING: &str = "DISABLE_AUTO_BATCHING";
|
const DISABLE_AUTO_BATCHING: &str = "DISABLE_AUTO_BATCHING";
|
||||||
|
|
||||||
const DEFAULT_LOG_EVERY_N: usize = 100000;
|
const DEFAULT_LOG_EVERY_N: usize = 100000;
|
||||||
|
|
||||||
#[derive(Debug, Clone, Parser, Serialize, Deserialize)]
|
#[derive(Debug, Clone, Parser, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "snake_case", deny_unknown_fields)]
|
||||||
pub struct IndexerOpts {
|
pub struct IndexerOpts {
|
||||||
/// The amount of documents to skip before printing
|
/// The amount of documents to skip before printing
|
||||||
/// a log regarding the indexing advancement.
|
/// a log regarding the indexing advancement.
|
||||||
@ -50,6 +50,7 @@ pub struct IndexerOpts {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, Parser, Default, Serialize, Deserialize)]
|
#[derive(Debug, Clone, Parser, Default, Serialize, Deserialize)]
|
||||||
|
#[serde(rename_all = "snake_case", deny_unknown_fields)]
|
||||||
pub struct SchedulerConfig {
|
pub struct SchedulerConfig {
|
||||||
/// The engine will disable task auto-batching,
|
/// The engine will disable task auto-batching,
|
||||||
/// and will sequencialy compute each task one by one.
|
/// and will sequencialy compute each task one by one.
|
||||||
@ -61,7 +62,13 @@ pub struct SchedulerConfig {
|
|||||||
impl IndexerOpts {
|
impl IndexerOpts {
|
||||||
/// Exports the values to their corresponding env vars if they are not set.
|
/// Exports the values to their corresponding env vars if they are not set.
|
||||||
pub fn export_to_env(self) {
|
pub fn export_to_env(self) {
|
||||||
if let Some(max_indexing_memory) = self.max_indexing_memory.0 {
|
let IndexerOpts {
|
||||||
|
max_indexing_memory,
|
||||||
|
max_indexing_threads,
|
||||||
|
log_every_n: _,
|
||||||
|
max_nb_chunks: _,
|
||||||
|
} = self;
|
||||||
|
if let Some(max_indexing_memory) = max_indexing_memory.0 {
|
||||||
export_to_env_if_not_present(
|
export_to_env_if_not_present(
|
||||||
MEILI_MAX_INDEXING_MEMORY,
|
MEILI_MAX_INDEXING_MEMORY,
|
||||||
max_indexing_memory.to_string(),
|
max_indexing_memory.to_string(),
|
||||||
@ -69,7 +76,7 @@ impl IndexerOpts {
|
|||||||
}
|
}
|
||||||
export_to_env_if_not_present(
|
export_to_env_if_not_present(
|
||||||
MEILI_MAX_INDEXING_THREADS,
|
MEILI_MAX_INDEXING_THREADS,
|
||||||
self.max_indexing_threads.0.to_string(),
|
max_indexing_threads.0.to_string(),
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -106,10 +113,10 @@ impl Default for IndexerOpts {
|
|||||||
|
|
||||||
impl SchedulerConfig {
|
impl SchedulerConfig {
|
||||||
pub fn export_to_env(self) {
|
pub fn export_to_env(self) {
|
||||||
export_to_env_if_not_present(
|
let SchedulerConfig {
|
||||||
DISABLE_AUTO_BATCHING,
|
disable_auto_batching,
|
||||||
self.disable_auto_batching.to_string(),
|
} = self;
|
||||||
);
|
export_to_env_if_not_present(DISABLE_AUTO_BATCHING, disable_auto_batching.to_string());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user