Fix typos

Found via `codespell -L crate,nam,hart,succeded`.
This commit is contained in:
Kian-Meng Ang 2022-09-20 22:39:35 +08:00
parent f4b81fa0a1
commit 740926e747
11 changed files with 21 additions and 21 deletions

View File

@ -85,7 +85,7 @@ get_latest() {
latest="" latest=""
current_tag="" current_tag=""
for release_info in $releases; do for release_info in $releases; do
if [ $i -eq 0 ]; then # Cheking tag_name if [ $i -eq 0 ]; then # Checking tag_name
if echo "$release_info" | grep -q "$GREP_SEMVER_REGEXP"; then # If it's not an alpha or beta release if echo "$release_info" | grep -q "$GREP_SEMVER_REGEXP"; then # If it's not an alpha or beta release
current_tag=$release_info current_tag=$release_info
else else

View File

@ -62,12 +62,12 @@ jobs:
- uses: actions/checkout@v3 - uses: actions/checkout@v3
- name: Download the issue template - name: Download the issue template
run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/roadmap-issue.md > $ISSUE_TEMPLATE run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/roadmap-issue.md > $ISSUE_TEMPLATE
- name: Replace all empty occurences in the templates - name: Replace all empty occurrences in the templates
run: | run: |
# Replace all <<version>> occurences # Replace all <<version>> occurrences
sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE
# Replace all <<milestone_id>> occurences # Replace all <<milestone_id>> occurrences
milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7) milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7)
sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE
@ -95,12 +95,12 @@ jobs:
- uses: actions/checkout@v3 - uses: actions/checkout@v3
- name: Download the issue template - name: Download the issue template
run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/changelog-issue.md > $ISSUE_TEMPLATE run: curl -s https://raw.githubusercontent.com/meilisearch/core-team/main/issue-templates/changelog-issue.md > $ISSUE_TEMPLATE
- name: Replace all empty occurences in the templates - name: Replace all empty occurrences in the templates
run: | run: |
# Replace all <<version>> occurences # Replace all <<version>> occurrences
sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE sed -i "s/<<version>>/$MILESTONE_VERSION/g" $ISSUE_TEMPLATE
# Replace all <<milestone_id>> occurences # Replace all <<milestone_id>> occurrences
milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7) milestone_id=$(echo $MILESTONE_URL | cut -d '/' -f 7)
sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE sed -i "s/<<milestone_id>>/$milestone_id/g" $ISSUE_TEMPLATE
- name: Create the issue - name: Create the issue

View File

@ -53,7 +53,7 @@ jobs:
uses: docker/metadata-action@v4 uses: docker/metadata-action@v4
with: with:
images: getmeili/meilisearch images: getmeili/meilisearch
# The lastest and `vX.Y` tags are only pushed for the official Meilisearch releases # The latest and `vX.Y` tags are only pushed for the official Meilisearch releases
# See https://github.com/docker/metadata-action#latest-tag # See https://github.com/docker/metadata-action#latest-tag
flavor: latest=false flavor: latest=false
tags: | tags: |

View File

@ -102,7 +102,7 @@ The full Meilisearch release process is described in [this guide](https://github
### Release assets ### Release assets
For each release, the following assets are created: For each release, the following assets are created:
- Binaries for differents platforms (Linux, MacOS, Windows and ARM architectures) are attached to the GitHub release - Binaries for different platforms (Linux, MacOS, Windows and ARM architectures) are attached to the GitHub release
- Binaries are pushed to HomeBrew and APT (not published for RC) - Binaries are pushed to HomeBrew and APT (not published for RC)
- Docker tags are created/updated: - Docker tags are created/updated:
- `vX.Y.Z` - `vX.Y.Z`

View File

@ -349,16 +349,16 @@ pub struct SearchAggregator {
// sort // sort
sort_with_geo_point: bool, sort_with_geo_point: bool,
// everytime a request has a filter, this field must be incremented by the number of terms it contains // every time a request has a filter, this field must be incremented by the number of terms it contains
sort_sum_of_criteria_terms: usize, sort_sum_of_criteria_terms: usize,
// everytime a request has a filter, this field must be incremented by one // every time a request has a filter, this field must be incremented by one
sort_total_number_of_criteria: usize, sort_total_number_of_criteria: usize,
// filter // filter
filter_with_geo_radius: bool, filter_with_geo_radius: bool,
// everytime a request has a filter, this field must be incremented by the number of terms it contains // every time a request has a filter, this field must be incremented by the number of terms it contains
filter_sum_of_criteria_terms: usize, filter_sum_of_criteria_terms: usize,
// everytime a request has a filter, this field must be incremented by one // every time a request has a filter, this field must be incremented by one
filter_total_number_of_criteria: usize, filter_total_number_of_criteria: usize,
used_syntax: HashMap<String, usize>, used_syntax: HashMap<String, usize>,
@ -366,7 +366,7 @@ pub struct SearchAggregator {
// The maximum number of terms in a q request // The maximum number of terms in a q request
max_terms_number: usize, max_terms_number: usize,
// everytime a search is done, we increment the counter linked to the used settings // every time a search is done, we increment the counter linked to the used settings
matching_strategy: HashMap<String, usize>, matching_strategy: HashMap<String, usize>,
// pagination // pagination

View File

@ -169,7 +169,7 @@ pub struct Opt {
} }
impl Opt { impl Opt {
/// Wether analytics should be enabled or not. /// Whether analytics should be enabled or not.
#[cfg(all(not(debug_assertions), feature = "analytics"))] #[cfg(all(not(debug_assertions), feature = "analytics"))]
pub fn analytics(&self) -> bool { pub fn analytics(&self) -> bool {
!self.no_analytics !self.no_analytics

View File

@ -145,7 +145,7 @@ pub fn error_code_from_str(s: &str) -> anyhow::Result<Code> {
"unsupported_media_type" => Code::UnsupportedMediaType, "unsupported_media_type" => Code::UnsupportedMediaType,
"dump_already_in_progress" => Code::DumpAlreadyInProgress, "dump_already_in_progress" => Code::DumpAlreadyInProgress,
"dump_process_failed" => Code::DumpProcessFailed, "dump_process_failed" => Code::DumpProcessFailed,
_ => bail!("unknow error code."), _ => bail!("unknown error code."),
}; };
Ok(code) Ok(code)

View File

@ -57,10 +57,10 @@ fn patch_updates(src: impl AsRef<Path>, dst: impl AsRef<Path>) -> anyhow::Result
let updates_path = src.as_ref().join("updates/data.jsonl"); let updates_path = src.as_ref().join("updates/data.jsonl");
let output_updates_path = dst.as_ref().join("updates/data.jsonl"); let output_updates_path = dst.as_ref().join("updates/data.jsonl");
create_dir_all(output_updates_path.parent().unwrap())?; create_dir_all(output_updates_path.parent().unwrap())?;
let udpates_file = File::open(updates_path)?; let updates_file = File::open(updates_path)?;
let mut output_update_file = File::create(output_updates_path)?; let mut output_update_file = File::create(output_updates_path)?;
serde_json::Deserializer::from_reader(udpates_file) serde_json::Deserializer::from_reader(updates_file)
.into_iter::<compat::v4::Task>() .into_iter::<compat::v4::Task>()
.try_for_each(|task| -> anyhow::Result<()> { .try_for_each(|task| -> anyhow::Result<()> {
let task: Task = task?.into(); let task: Task = task?.into();

View File

@ -27,7 +27,7 @@ pub const DEFAULT_CROP_MARKER: fn() -> String = || "…".to_string();
pub const DEFAULT_HIGHLIGHT_PRE_TAG: fn() -> String = || "<em>".to_string(); pub const DEFAULT_HIGHLIGHT_PRE_TAG: fn() -> String = || "<em>".to_string();
pub const DEFAULT_HIGHLIGHT_POST_TAG: fn() -> String = || "</em>".to_string(); pub const DEFAULT_HIGHLIGHT_POST_TAG: fn() -> String = || "</em>".to_string();
/// The maximimum number of results that the engine /// The maximum number of results that the engine
/// will be able to return in one search call. /// will be able to return in one search call.
pub const DEFAULT_PAGINATION_MAX_TOTAL_HITS: usize = 1000; pub const DEFAULT_PAGINATION_MAX_TOTAL_HITS: usize = 1000;

View File

@ -51,7 +51,7 @@ impl MapIndexStore {
#[async_trait::async_trait] #[async_trait::async_trait]
impl IndexStore for MapIndexStore { impl IndexStore for MapIndexStore {
async fn create(&self, uuid: Uuid) -> Result<Index> { async fn create(&self, uuid: Uuid) -> Result<Index> {
// We need to keep the lock until we are sure the db file has been opened correclty, to // We need to keep the lock until we are sure the db file has been opened correctly, to
// ensure that another db is not created at the same time. // ensure that another db is not created at the same time.
let mut lock = self.index_store.write().await; let mut lock = self.index_store.write().await;

View File

@ -63,7 +63,7 @@ impl Store {
/// Returns the id for the next task. /// Returns the id for the next task.
/// ///
/// The required `mut txn` acts as a reservation system. It guarantees that as long as you commit /// The required `mut txn` acts as a reservation system. It guarantees that as long as you commit
/// the task to the store in the same transaction, no one else will hav this task id. /// the task to the store in the same transaction, no one else will have this task id.
pub fn next_task_id(&self, txn: &mut RwTxn) -> Result<TaskId> { pub fn next_task_id(&self, txn: &mut RwTxn) -> Result<TaskId> {
let id = self let id = self
.tasks .tasks