mirror of
https://github.com/meilisearch/MeiliSearch
synced 2024-11-22 21:04:27 +01:00
Fix typos
Found via `codespell -L crate,nam,hart,succeded`.
This commit is contained in:
parent
f4b81fa0a1
commit
740926e747
2
.github/scripts/is-latest-release.sh
vendored
2
.github/scripts/is-latest-release.sh
vendored
@ -85,7 +85,7 @@ get_latest() {
|
|||||||
latest=""
|
latest=""
|
||||||
current_tag=""
|
current_tag=""
|
||||||
for release_info in $releases; do
|
for release_info in $releases; do
|
||||||
if [ $i -eq 0 ]; then # Cheking tag_name
|
if [ $i -eq 0 ]; then # Checking tag_name
|
||||||
if echo "$release_info" | grep -q "$GREP_SEMVER_REGEXP"; then # If it's not an alpha or beta release
|
if echo "$release_info" | grep -q "$GREP_SEMVER_REGEXP"; then # If it's not an alpha or beta release
|
||||||
current_tag=$release_info
|
current_tag=$release_info
|
||||||
else
|
else
|
||||||
|
12
.github/workflows/milestone-workflow.yml
vendored
12
.github/workflows/milestone-workflow.yml
vendored
@ -62,12 +62,12 @@ jobs:
|
|||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v3
|
||||||
- name: Download the issue template
|
- name: Download the issue template
|
||||||
run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/roadmap-issue.md > $ISSUE_TEMPLATE
|
run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/roadmap-issue.md > $ISSUE_TEMPLATE
|
||||||
- name: Replace all empty occurences in the templates
|
- name: Replace all empty occurrences in the templates
|
||||||
run: |
|
run: |
|
||||||
# Replace all <<version>> occurences
|
# Replace all <<version>> occurrences
|
||||||
sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE
|
sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE
|
||||||
|
|
||||||
# Replace all <<milestone_id>> occurences
|
# Replace all <<milestone_id>> occurrences
|
||||||
milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7)
|
milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7)
|
||||||
sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE
|
sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE
|
||||||
|
|
||||||
@ -95,12 +95,12 @@ jobs:
|
|||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v3
|
||||||
- name: Download the issue template
|
- name: Download the issue template
|
||||||
run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/changelog-issue.md > $ISSUE_TEMPLATE
|
run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/changelog-issue.md > $ISSUE_TEMPLATE
|
||||||
- name: Replace all empty occurences in the templates
|
- name: Replace all empty occurrences in the templates
|
||||||
run: |
|
run: |
|
||||||
# Replace all <<version>> occurences
|
# Replace all <<version>> occurrences
|
||||||
sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE
|
sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE
|
||||||
|
|
||||||
# Replace all <<milestone_id>> occurences
|
# Replace all <<milestone_id>> occurrences
|
||||||
milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7)
|
milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7)
|
||||||
sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE
|
sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE
|
||||||
- name: Create the issue
|
- name: Create the issue
|
||||||
|
2
.github/workflows/publish-docker-images.yml
vendored
2
.github/workflows/publish-docker-images.yml
vendored
@ -53,7 +53,7 @@ jobs:
|
|||||||
uses: docker/metadata-action@v4
|
uses: docker/metadata-action@v4
|
||||||
with:
|
with:
|
||||||
images: getmeili/meilisearch
|
images: getmeili/meilisearch
|
||||||
# The lastest and `vX.Y` tags are only pushed for the official Meilisearch releases
|
# The latest and `vX.Y` tags are only pushed for the official Meilisearch releases
|
||||||
# See https://github.com/docker/metadata-action#latest-tag
|
# See https://github.com/docker/metadata-action#latest-tag
|
||||||
flavor: latest=false
|
flavor: latest=false
|
||||||
tags: |
|
tags: |
|
||||||
|
@ -102,7 +102,7 @@ The full Meilisearch release process is described in [this guide](https://github
|
|||||||
### Release assets
|
### Release assets
|
||||||
|
|
||||||
For each release, the following assets are created:
|
For each release, the following assets are created:
|
||||||
- Binaries for differents platforms (Linux, MacOS, Windows and ARM architectures) are attached to the GitHub release
|
- Binaries for different platforms (Linux, MacOS, Windows and ARM architectures) are attached to the GitHub release
|
||||||
- Binaries are pushed to HomeBrew and APT (not published for RC)
|
- Binaries are pushed to HomeBrew and APT (not published for RC)
|
||||||
- Docker tags are created/updated:
|
- Docker tags are created/updated:
|
||||||
- `vX.Y.Z`
|
- `vX.Y.Z`
|
||||||
|
@ -349,16 +349,16 @@ pub struct SearchAggregator {
|
|||||||
|
|
||||||
// sort
|
// sort
|
||||||
sort_with_geo_point: bool,
|
sort_with_geo_point: bool,
|
||||||
// everytime a request has a filter, this field must be incremented by the number of terms it contains
|
// every time a request has a filter, this field must be incremented by the number of terms it contains
|
||||||
sort_sum_of_criteria_terms: usize,
|
sort_sum_of_criteria_terms: usize,
|
||||||
// everytime a request has a filter, this field must be incremented by one
|
// every time a request has a filter, this field must be incremented by one
|
||||||
sort_total_number_of_criteria: usize,
|
sort_total_number_of_criteria: usize,
|
||||||
|
|
||||||
// filter
|
// filter
|
||||||
filter_with_geo_radius: bool,
|
filter_with_geo_radius: bool,
|
||||||
// everytime a request has a filter, this field must be incremented by the number of terms it contains
|
// every time a request has a filter, this field must be incremented by the number of terms it contains
|
||||||
filter_sum_of_criteria_terms: usize,
|
filter_sum_of_criteria_terms: usize,
|
||||||
// everytime a request has a filter, this field must be incremented by one
|
// every time a request has a filter, this field must be incremented by one
|
||||||
filter_total_number_of_criteria: usize,
|
filter_total_number_of_criteria: usize,
|
||||||
used_syntax: HashMap<String, usize>,
|
used_syntax: HashMap<String, usize>,
|
||||||
|
|
||||||
@ -366,7 +366,7 @@ pub struct SearchAggregator {
|
|||||||
// The maximum number of terms in a q request
|
// The maximum number of terms in a q request
|
||||||
max_terms_number: usize,
|
max_terms_number: usize,
|
||||||
|
|
||||||
// everytime a search is done, we increment the counter linked to the used settings
|
// every time a search is done, we increment the counter linked to the used settings
|
||||||
matching_strategy: HashMap<String, usize>,
|
matching_strategy: HashMap<String, usize>,
|
||||||
|
|
||||||
// pagination
|
// pagination
|
||||||
|
@ -169,7 +169,7 @@ pub struct Opt {
|
|||||||
}
|
}
|
||||||
|
|
||||||
impl Opt {
|
impl Opt {
|
||||||
/// Wether analytics should be enabled or not.
|
/// Whether analytics should be enabled or not.
|
||||||
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
#[cfg(all(not(debug_assertions), feature = "analytics"))]
|
||||||
pub fn analytics(&self) -> bool {
|
pub fn analytics(&self) -> bool {
|
||||||
!self.no_analytics
|
!self.no_analytics
|
||||||
|
@ -145,7 +145,7 @@ pub fn error_code_from_str(s: &str) -> anyhow::Result<Code> {
|
|||||||
"unsupported_media_type" => Code::UnsupportedMediaType,
|
"unsupported_media_type" => Code::UnsupportedMediaType,
|
||||||
"dump_already_in_progress" => Code::DumpAlreadyInProgress,
|
"dump_already_in_progress" => Code::DumpAlreadyInProgress,
|
||||||
"dump_process_failed" => Code::DumpProcessFailed,
|
"dump_process_failed" => Code::DumpProcessFailed,
|
||||||
_ => bail!("unknow error code."),
|
_ => bail!("unknown error code."),
|
||||||
};
|
};
|
||||||
|
|
||||||
Ok(code)
|
Ok(code)
|
||||||
|
@ -57,10 +57,10 @@ fn patch_updates(src: impl AsRef<Path>, dst: impl AsRef<Path>) -> anyhow::Result
|
|||||||
let updates_path = src.as_ref().join("updates/data.jsonl");
|
let updates_path = src.as_ref().join("updates/data.jsonl");
|
||||||
let output_updates_path = dst.as_ref().join("updates/data.jsonl");
|
let output_updates_path = dst.as_ref().join("updates/data.jsonl");
|
||||||
create_dir_all(output_updates_path.parent().unwrap())?;
|
create_dir_all(output_updates_path.parent().unwrap())?;
|
||||||
let udpates_file = File::open(updates_path)?;
|
let updates_file = File::open(updates_path)?;
|
||||||
let mut output_update_file = File::create(output_updates_path)?;
|
let mut output_update_file = File::create(output_updates_path)?;
|
||||||
|
|
||||||
serde_json::Deserializer::from_reader(udpates_file)
|
serde_json::Deserializer::from_reader(updates_file)
|
||||||
.into_iter::<compat::v4::Task>()
|
.into_iter::<compat::v4::Task>()
|
||||||
.try_for_each(|task| -> anyhow::Result<()> {
|
.try_for_each(|task| -> anyhow::Result<()> {
|
||||||
let task: Task = task?.into();
|
let task: Task = task?.into();
|
||||||
|
@ -27,7 +27,7 @@ pub const DEFAULT_CROP_MARKER: fn() -> String = || "…".to_string();
|
|||||||
pub const DEFAULT_HIGHLIGHT_PRE_TAG: fn() -> String = || "<em>".to_string();
|
pub const DEFAULT_HIGHLIGHT_PRE_TAG: fn() -> String = || "<em>".to_string();
|
||||||
pub const DEFAULT_HIGHLIGHT_POST_TAG: fn() -> String = || "</em>".to_string();
|
pub const DEFAULT_HIGHLIGHT_POST_TAG: fn() -> String = || "</em>".to_string();
|
||||||
|
|
||||||
/// The maximimum number of results that the engine
|
/// The maximum number of results that the engine
|
||||||
/// will be able to return in one search call.
|
/// will be able to return in one search call.
|
||||||
pub const DEFAULT_PAGINATION_MAX_TOTAL_HITS: usize = 1000;
|
pub const DEFAULT_PAGINATION_MAX_TOTAL_HITS: usize = 1000;
|
||||||
|
|
||||||
|
@ -51,7 +51,7 @@ impl MapIndexStore {
|
|||||||
#[async_trait::async_trait]
|
#[async_trait::async_trait]
|
||||||
impl IndexStore for MapIndexStore {
|
impl IndexStore for MapIndexStore {
|
||||||
async fn create(&self, uuid: Uuid) -> Result<Index> {
|
async fn create(&self, uuid: Uuid) -> Result<Index> {
|
||||||
// We need to keep the lock until we are sure the db file has been opened correclty, to
|
// We need to keep the lock until we are sure the db file has been opened correctly, to
|
||||||
// ensure that another db is not created at the same time.
|
// ensure that another db is not created at the same time.
|
||||||
let mut lock = self.index_store.write().await;
|
let mut lock = self.index_store.write().await;
|
||||||
|
|
||||||
|
@ -63,7 +63,7 @@ impl Store {
|
|||||||
/// Returns the id for the next task.
|
/// Returns the id for the next task.
|
||||||
///
|
///
|
||||||
/// The required `mut txn` acts as a reservation system. It guarantees that as long as you commit
|
/// The required `mut txn` acts as a reservation system. It guarantees that as long as you commit
|
||||||
/// the task to the store in the same transaction, no one else will hav this task id.
|
/// the task to the store in the same transaction, no one else will have this task id.
|
||||||
pub fn next_task_id(&self, txn: &mut RwTxn) -> Result<TaskId> {
|
pub fn next_task_id(&self, txn: &mut RwTxn) -> Result<TaskId> {
|
||||||
let id = self
|
let id = self
|
||||||
.tasks
|
.tasks
|
||||||
|
Loading…
Reference in New Issue
Block a user