From dd4bc569b9f73e8c2445bed90d82c92d98aada03 Mon Sep 17 00:00:00 2001 From: Andrea Gelmini Date: Wed, 10 Jun 2020 06:24:09 +0200 Subject: Fix typos Correct various typos in the comments and tests. Reviewed-by: Ryan Moeller Reviewed-by: Matthew Ahrens Reviewed-by: Brian Behlendorf Signed-off-by: Andrea Gelmini Closes #10423 --- module/zfs/abd.c | 6 +++--- module/zfs/arc.c | 4 ++-- module/zfs/btree.c | 10 +++++----- module/zfs/dsl_bookmark.c | 2 +- module/zfs/metaslab.c | 2 +- module/zfs/vdev_raidz.c | 4 ++-- module/zfs/vdev_raidz_math_powerpc_altivec.c | 2 +- module/zfs/zfs_ioctl.c | 2 +- module/zfs/zvol.c | 2 +- 9 files changed, 17 insertions(+), 17 deletions(-) (limited to 'module/zfs') diff --git a/module/zfs/abd.c b/module/zfs/abd.c index a3e58ebc5..269f2a7e5 100644 --- a/module/zfs/abd.c +++ b/module/zfs/abd.c @@ -388,7 +388,7 @@ abd_gang_add(abd_t *pabd, abd_t *cabd, boolean_t free_on_free) /* * If the child ABD is already part of another * gang ABD then we must allocate a new - * ABD to use a seperate link. We mark the newly + * ABD to use a separate link. We mark the newly * allocated ABD with ABD_FLAG_GANG_FREE, before * adding it to the gang ABD's list, to make the * gang ABD aware that it is responsible to call @@ -397,7 +397,7 @@ abd_gang_add(abd_t *pabd, abd_t *cabd, boolean_t free_on_free) * data over into the newly allocated ABD. * * An ABD may become part of multiple gang ABD's. For - * example, when writting ditto bocks, the same ABD + * example, when writing ditto bocks, the same ABD * is used to write 2 or 3 locations with 2 or 3 * zio_t's. Each of the zio's may be aggregated with * different adjacent zio's. zio aggregation uses gang @@ -406,7 +406,7 @@ abd_gang_add(abd_t *pabd, abd_t *cabd, boolean_t free_on_free) * * The ASSERT below is to make sure that if * free_on_free is passed as B_TRUE, the ABD can - * not be in mulitple gang ABD's. The gang ABD + * not be in multiple gang ABD's. The gang ABD * can not be responsible for cleaning up the child * ABD memory allocation if the ABD can be in * multiple gang ABD's at one time. diff --git a/module/zfs/arc.c b/module/zfs/arc.c index e7ad976af..c2e6720d3 100644 --- a/module/zfs/arc.c +++ b/module/zfs/arc.c @@ -922,7 +922,7 @@ static void l2arc_hdr_restore(const l2arc_log_ent_phys_t *le, static void l2arc_log_blk_commit(l2arc_dev_t *dev, zio_t *pio, l2arc_write_callback_t *cb); -/* L2ARC persistence auxilliary routines. */ +/* L2ARC persistence auxiliary routines. */ boolean_t l2arc_log_blkptr_valid(l2arc_dev_t *dev, const l2arc_log_blkptr_t *lbp); static boolean_t l2arc_log_blk_insert(l2arc_dev_t *dev, @@ -8382,7 +8382,7 @@ top: rerun = B_FALSE; if (dev->l2ad_hand >= (dev->l2ad_end - distance)) { /* - * When there is no space to accomodate upcoming writes, + * When there is no space to accommodate upcoming writes, * evict to the end. Then bump the write and evict hands * to the start and iterate. This iteration does not * happen indefinitely as we make sure in diff --git a/module/zfs/btree.c b/module/zfs/btree.c index da1886a57..d9b908dd2 100644 --- a/module/zfs/btree.c +++ b/module/zfs/btree.c @@ -342,7 +342,7 @@ zfs_btree_find(zfs_btree_t *tree, const void *value, zfs_btree_index_t *where) * * / * * * *\ * * * ... * * * * --------------- * - * Note that a parellelogram shift is always shaped like a "left-leaning" + * Note that a parallelogram shift is always shaped like a "left-leaning" * parallelogram, where the starting index of the children being moved is * always one higher than the starting index of the elements being moved. No * "right-leaning" parallelogram shifts are needed (shifts where the starting @@ -578,7 +578,7 @@ zfs_btree_insert_into_parent(zfs_btree_t *tree, zfs_btree_hdr_t *old_node, ASSERT3P(parent->btc_children[offset], ==, old_node); /* - * If the parent isn't full, shift things to accomodate our insertions + * If the parent isn't full, shift things to accommodate our insertions * and return. */ if (par_hdr->bth_count != BTREE_CORE_ELEMS) { @@ -651,7 +651,7 @@ zfs_btree_insert_into_parent(zfs_btree_t *tree, zfs_btree_hdr_t *old_node, /* * Move the new separator into the right half, and replace it * with buf. We also need to shift back the elements in the - * right half to accomodate new_node. + * right half to accommodate new_node. */ bt_shift_core_right(tree, new_parent, 0, move_count, BSS_TRAPEZOID); @@ -1366,7 +1366,7 @@ zfs_btree_remove_from_node(zfs_btree_t *tree, zfs_btree_core_t *node, /* * Now we try to take a node from a neighbor. We check left, then * right. If the neighbor exists and has more than the minimum number - * of elements, we move the separator betweeen us and them to our + * of elements, we move the separator between us and them to our * node, move their closest element (last for left, first for right) * to the separator, and move their closest child to our node. Along * the way we need to collapse the gap made by idx, and (for our right @@ -1625,7 +1625,7 @@ zfs_btree_remove_idx(zfs_btree_t *tree, zfs_btree_index_t *where) /* * Now we try to take a node from a sibling. We check left, then * right. If they exist and have more than the minimum number of - * elements, we move the separator betweeen us and them to our node + * elements, we move the separator between us and them to our node * and move their closest element (last for left, first for right) to * the separator. Along the way we need to collapse the gap made by * idx, and (for our right neighbor) the gap made by removing their diff --git a/module/zfs/dsl_bookmark.c b/module/zfs/dsl_bookmark.c index 7e6a5ceaa..16bf2c441 100644 --- a/module/zfs/dsl_bookmark.c +++ b/module/zfs/dsl_bookmark.c @@ -123,7 +123,7 @@ dsl_bookmark_lookup(dsl_pool_t *dp, const char *fullname, /* * Validates that * - bmark is a full dataset path of a bookmark (bookmark_namecheck) - * - source is a full path of a snaphot or bookmark + * - source is a full path of a snapshot or bookmark * ({bookmark,snapshot}_namecheck) * * Returns 0 if valid, -1 otherwise. diff --git a/module/zfs/metaslab.c b/module/zfs/metaslab.c index 1fc44399f..549055628 100644 --- a/module/zfs/metaslab.c +++ b/module/zfs/metaslab.c @@ -3547,7 +3547,7 @@ metaslab_condense(metaslab_t *msp, dmu_tx_t *tx) * 4] At this point, we would ideally like to add all segments * in the ms_allocatable tree from the condense tree. This way * we would write all the entries of the condense tree as the - * condensed space map, which would only contain freeed + * condensed space map, which would only contain freed * segments with everything else assumed to be allocated. * * Doing so can be prohibitively expensive as ms_allocatable can diff --git a/module/zfs/vdev_raidz.c b/module/zfs/vdev_raidz.c index 3408ddf72..be3466673 100644 --- a/module/zfs/vdev_raidz.c +++ b/module/zfs/vdev_raidz.c @@ -982,7 +982,7 @@ vdev_raidz_reconstruct_pq(raidz_map_t *rm, int *tgts, int ntgts) /* BEGIN CSTYLED */ /* * In the general case of reconstruction, we must solve the system of linear - * equations defined by the coeffecients used to generate parity as well as + * equations defined by the coefficients used to generate parity as well as * the contents of the data and parity disks. This can be expressed with * vectors for the original data (D) and the actual data (d) and parity (p) * and a matrix composed of the identity matrix (I) and a dispersal matrix (V): @@ -996,7 +996,7 @@ vdev_raidz_reconstruct_pq(raidz_map_t *rm, int *tgts, int ntgts) * ~~ ~~ ~~ ~~ * * I is simply a square identity matrix of size n, and V is a vandermonde - * matrix defined by the coeffecients we chose for the various parity columns + * matrix defined by the coefficients we chose for the various parity columns * (1, 2, 4). Note that these values were chosen both for simplicity, speedy * computation as well as linear separability. * diff --git a/module/zfs/vdev_raidz_math_powerpc_altivec.c b/module/zfs/vdev_raidz_math_powerpc_altivec.c index feabfa5c0..88f2646ae 100644 --- a/module/zfs/vdev_raidz_math_powerpc_altivec.c +++ b/module/zfs/vdev_raidz_math_powerpc_altivec.c @@ -4333,5 +4333,5 @@ __attribute__((aligned(256))) gf_clmul_mod_lt[4*256][16] = { 0xf8, 0x07, 0x06, 0xf9, 0x04, 0xfb, 0xfa, 0x05 } }; /* END CSTYLED */ -#endif // ENDIANESS +#endif // ENDIANNESS #endif /* defined(__powerpc__) */ diff --git a/module/zfs/zfs_ioctl.c b/module/zfs/zfs_ioctl.c index 8b6dca18c..33e1620ec 100644 --- a/module/zfs/zfs_ioctl.c +++ b/module/zfs/zfs_ioctl.c @@ -3520,7 +3520,7 @@ zfs_ioc_log_history(const char *unused, nvlist_t *innvl, nvlist_t *outnvl) * of the grubenv file. The file is stored as raw ASCII, and is protected by * an embedded checksum. By default, GRUB will check if the boot filesystem * supports storing the environment data in a special location, and if so, - * will invoke filesystem specific logic to retrieve it. This can be overriden + * will invoke filesystem specific logic to retrieve it. This can be overridden * by a variable, should the user so desire. */ /* ARGSUSED */ diff --git a/module/zfs/zvol.c b/module/zfs/zvol.c index bbda652ee..2b20b02e4 100644 --- a/module/zfs/zvol.c +++ b/module/zfs/zvol.c @@ -1717,7 +1717,7 @@ zvol_fini_impl(void) /* * The call to "zvol_remove_minors_impl" may dispatch entries to - * the system_taskq, but it doesn't wait for those entires to + * the system_taskq, but it doesn't wait for those entries to * complete before it returns. Thus, we must wait for all of the * removals to finish, before we can continue. */ -- cgit v1.2.3