summaryrefslogtreecommitdiffstats
path: root/module/zfs
diff options
context:
space:
mode:
authorAndrea Gelmini <[email protected]>2021-04-02 18:38:53 -0700
committerBrian Behlendorf <[email protected]>2021-04-07 13:27:11 -0700
commitca7af7f675bb93bc6610437973e3a59c8853991e (patch)
treec34c18ef6df827d6a0550e2664cf1dc56f42cbdf /module/zfs
parent35cce6ea6375c8f2e36d4569776f837fcac64289 (diff)
Fix various typos
Correct an assortment of typos throughout the code base. Reviewed-by: Brian Behlendorf <[email protected]> Reviewed-by: Matthew Ahrens <[email protected]> Reviewed-by: Ryan Moeller <[email protected]> Signed-off-by: Andrea Gelmini <[email protected]> Closes #11774
Diffstat (limited to 'module/zfs')
-rw-r--r--module/zfs/abd.c2
-rw-r--r--module/zfs/arc.c2
-rw-r--r--module/zfs/dsl_bookmark.c6
-rw-r--r--module/zfs/spa.c2
-rw-r--r--module/zfs/vdev.c8
-rw-r--r--module/zfs/vdev_raidz.c2
-rw-r--r--module/zfs/vdev_rebuild.c2
-rw-r--r--module/zfs/zfs_ioctl.c2
8 files changed, 13 insertions, 13 deletions
diff --git a/module/zfs/abd.c b/module/zfs/abd.c
index 1e6645c90..2d1be9752 100644
--- a/module/zfs/abd.c
+++ b/module/zfs/abd.c
@@ -381,7 +381,7 @@ abd_gang_add_gang(abd_t *pabd, abd_t *cabd, boolean_t free_on_free)
child = list_next(&ABD_GANG(cabd).abd_gang_chain, child)) {
/*
* We always pass B_FALSE for free_on_free as it is the
- * original child gang ABDs responsibilty to determine
+ * original child gang ABDs responsibility to determine
* if any of its child ABDs should be free'd on the call
* to abd_free().
*/
diff --git a/module/zfs/arc.c b/module/zfs/arc.c
index 55c71a382..f0ae3938a 100644
--- a/module/zfs/arc.c
+++ b/module/zfs/arc.c
@@ -5036,7 +5036,7 @@ arc_reap_cb(void *arg, zthr_t *zthr)
* memory in the system at a fraction of the arc_size (1/128th by
* default). If oversubscribed (free_memory < 0) then reduce the
* target arc_size by the deficit amount plus the fractional
- * amount. If free memory is positive but less then the fractional
+ * amount. If free memory is positive but less than the fractional
* amount, reduce by what is needed to hit the fractional amount.
*/
free_memory = arc_available_memory();
diff --git a/module/zfs/dsl_bookmark.c b/module/zfs/dsl_bookmark.c
index 2faf1af52..bead7da22 100644
--- a/module/zfs/dsl_bookmark.c
+++ b/module/zfs/dsl_bookmark.c
@@ -236,7 +236,7 @@ dsl_bookmark_create_check_impl(dsl_pool_t *dp,
error = SET_ERROR(EEXIST);
goto eholdnewbmds;
default:
- /* dsl_bookmark_lookup_impl already did SET_ERRROR */
+ /* dsl_bookmark_lookup_impl already did SET_ERROR */
goto eholdnewbmds;
}
@@ -271,7 +271,7 @@ dsl_bookmark_create_check_impl(dsl_pool_t *dp,
error = SET_ERROR(ZFS_ERR_BOOKMARK_SOURCE_NOT_ANCESTOR);
break;
default:
- /* dsl_bookmark_lookup already did SET_ERRROR */
+ /* dsl_bookmark_lookup already did SET_ERROR */
break;
}
} else {
@@ -536,7 +536,7 @@ dsl_bookmark_create_sync_impl_book(
* Reasoning:
* - The zbm_redaction_obj would be referred to by both source and new
* bookmark, but would be destroyed once either source or new is
- * destroyed, resulting in use-after-free of the referrred object.
+ * destroyed, resulting in use-after-free of the referred object.
* - User expectation when issuing the `zfs bookmark` command is that
* a normal bookmark of the source is created
*
diff --git a/module/zfs/spa.c b/module/zfs/spa.c
index 5170c9ca2..73fa1809b 100644
--- a/module/zfs/spa.c
+++ b/module/zfs/spa.c
@@ -6496,7 +6496,7 @@ spa_vdev_add(spa_t *spa, nvlist_t *nvroot)
/*
* The virtual dRAID spares must be added after vdev tree is created
- * and the vdev guids are generated. The guid of their assoicated
+ * and the vdev guids are generated. The guid of their associated
* dRAID is stored in the config and used when opening the spare.
*/
if ((error = vdev_draid_spare_create(nvroot, vd, &ndraid,
diff --git a/module/zfs/vdev.c b/module/zfs/vdev.c
index c536a1c6c..69e44b48e 100644
--- a/module/zfs/vdev.c
+++ b/module/zfs/vdev.c
@@ -1372,7 +1372,7 @@ vdev_metaslab_group_create(vdev_t *vd)
/*
* The spa ashift min/max only apply for the normal metaslab
- * class. Class destination is late binding so ashift boundry
+ * class. Class destination is late binding so ashift boundary
* setting had to wait until now.
*/
if (vd->vdev_top == vd && vd->vdev_ashift != 0 &&
@@ -2046,7 +2046,7 @@ vdev_open(vdev_t *vd)
vd->vdev_max_asize = max_asize;
/*
- * If the vdev_ashift was not overriden at creation time,
+ * If the vdev_ashift was not overridden at creation time,
* then set it the logical ashift and optimize the ashift.
*/
if (vd->vdev_ashift == 0) {
@@ -2116,7 +2116,7 @@ vdev_open(vdev_t *vd)
}
/*
- * Track the the minimum allocation size.
+ * Track the minimum allocation size.
*/
if (vd->vdev_top == vd && vd->vdev_ashift != 0 &&
vd->vdev_islog == 0 && vd->vdev_aux == NULL) {
@@ -4570,7 +4570,7 @@ vdev_stat_update(zio_t *zio, uint64_t psize)
/*
* Solely for the purposes of 'zpool iostat -lqrw'
- * reporting use the priority to catagorize the IO.
+ * reporting use the priority to categorize the IO.
* Only the following are reported to user space:
*
* ZIO_PRIORITY_SYNC_READ,
diff --git a/module/zfs/vdev_raidz.c b/module/zfs/vdev_raidz.c
index db753ec16..020b3bc95 100644
--- a/module/zfs/vdev_raidz.c
+++ b/module/zfs/vdev_raidz.c
@@ -1984,7 +1984,7 @@ raidz_reconstruct(zio_t *zio, int *ltgts, int ntgts, int nparity)
* 2 4 5 first: increment to 3
* 3 4 5 done
*
- * This strategy works for dRAID but is less effecient when there are a large
+ * This strategy works for dRAID but is less efficient when there are a large
* number of child vdevs and therefore permutations to check. Furthermore,
* since the raidz_map_t rows likely do not overlap reconstruction would be
* possible as long as there are no more than nparity data errors per row.
diff --git a/module/zfs/vdev_rebuild.c b/module/zfs/vdev_rebuild.c
index a77ff99fa..aa79642aa 100644
--- a/module/zfs/vdev_rebuild.c
+++ b/module/zfs/vdev_rebuild.c
@@ -81,7 +81,7 @@
* Advantages:
*
* - Sequential reconstruction is performed in LBA order which may be faster
- * than healing reconstruction particularly when using using HDDs (or
+ * than healing reconstruction particularly when using HDDs (or
* especially with SMR devices). Only allocated capacity is resilvered.
*
* - Sequential reconstruction is not constrained by ZFS block boundaries.
diff --git a/module/zfs/zfs_ioctl.c b/module/zfs/zfs_ioctl.c
index 5f291d067..7f929df16 100644
--- a/module/zfs/zfs_ioctl.c
+++ b/module/zfs/zfs_ioctl.c
@@ -233,7 +233,7 @@ unsigned long zfs_max_nvlist_src_size = 0;
/*
* When logging the output nvlist of an ioctl in the on-disk history, limit
- * the logged size to this many bytes. This must be less then DMU_MAX_ACCESS.
+ * the logged size to this many bytes. This must be less than DMU_MAX_ACCESS.
* This applies primarily to zfs_ioc_channel_program().
*/
unsigned long zfs_history_output_max = 1024 * 1024;