aboutsummaryrefslogtreecommitdiffstats
path: root/tests/zfs-tests
diff options
context:
space:
mode:
authorAndrea Gelmini <[email protected]>2019-09-03 03:14:53 +0200
committerBrian Behlendorf <[email protected]>2019-09-02 18:14:53 -0700
commitc6e457dffb24bdc3efaea73b99480d2cf6567128 (patch)
tree030468c138d6ac4864bc89df93ba2d0656cef2b3 /tests/zfs-tests
parentcb14aa4ca9a2ba84beeb2165486e65c441272c04 (diff)
Fix typos in tests/
Reviewed-by: Ryan Moeller <[email protected]> Reviewed-by: Richard Laager <[email protected]> Reviewed-by: Brian Behlendorf <[email protected]> Signed-off-by: Andrea Gelmini <[email protected]> Closes #9250
Diffstat (limited to 'tests/zfs-tests')
-rw-r--r--tests/zfs-tests/cmd/mmapwrite/mmapwrite.c2
-rw-r--r--tests/zfs-tests/include/blkdev.shlib2
-rw-r--r--tests/zfs-tests/include/libtest.shlib6
-rw-r--r--tests/zfs-tests/tests/functional/acl/acl_common.kshlib12
-rwxr-xr-xtests/zfs-tests/tests/functional/cachefile/cachefile_004_pos.ksh4
-rwxr-xr-xtests/zfs-tests/tests/functional/casenorm/insensitive_formd_lookup.ksh2
-rwxr-xr-xtests/zfs-tests/tests/functional/channel_program/synctask_core/tst.list_user_props.ksh2
-rwxr-xr-xtests/zfs-tests/tests/functional/channel_program/synctask_core/tst.terminate_by_signal.ksh2
-rwxr-xr-xtests/zfs-tests/tests/functional/cli_root/zfs_clone/zfs_clone_010_pos.ksh2
-rwxr-xr-xtests/zfs-tests/tests/functional/cli_root/zfs_copies/zfs_copies_002_pos.ksh2
-rwxr-xr-xtests/zfs-tests/tests/functional/cli_root/zfs_create/zfs_create_011_pos.ksh2
-rwxr-xr-xtests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_001_pos.ksh2
-rwxr-xr-xtests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_005_neg.ksh4
-rwxr-xr-xtests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_014_pos.ksh4
14 files changed, 24 insertions, 24 deletions
diff --git a/tests/zfs-tests/cmd/mmapwrite/mmapwrite.c b/tests/zfs-tests/cmd/mmapwrite/mmapwrite.c
index b9915d5d3..458d6d8e4 100644
--- a/tests/zfs-tests/cmd/mmapwrite/mmapwrite.c
+++ b/tests/zfs-tests/cmd/mmapwrite/mmapwrite.c
@@ -43,7 +43,7 @@
* is hold) occurred, zfs_dirty_inode open a txg failed, and wait previous
* txg "n" completed.
* 3. context #1 call uiomove to write, however page fault is occurred in
- * uiomove, which means it need mm_sem, but mm_sem is hold by
+ * uiomove, which means it needs mm_sem, but mm_sem is hold by
* context #2, so it stuck and can't complete, then txg "n" will not
* complete.
*
diff --git a/tests/zfs-tests/include/blkdev.shlib b/tests/zfs-tests/include/blkdev.shlib
index ca8807e82..87500e92a 100644
--- a/tests/zfs-tests/include/blkdev.shlib
+++ b/tests/zfs-tests/include/blkdev.shlib
@@ -131,7 +131,7 @@ function is_loop_device #disk
}
#
-# Check if the given device is a multipath device and if there is a sybolic
+# Check if the given device is a multipath device and if there is a symbolic
# link to a device mapper and to a disk
# Currently no support for dm devices alone without multipath
#
diff --git a/tests/zfs-tests/include/libtest.shlib b/tests/zfs-tests/include/libtest.shlib
index cda4b04cd..ed68cac06 100644
--- a/tests/zfs-tests/include/libtest.shlib
+++ b/tests/zfs-tests/include/libtest.shlib
@@ -869,7 +869,7 @@ function zero_partitions #<whole_disk_name>
# Size should be specified with units as per
# the `format` command requirements eg. 100mb 3gb
#
-# NOTE: This entire interface is problematic for the Linux parted utilty
+# NOTE: This entire interface is problematic for the Linux parted utility
# which requires the end of the partition to be specified. It would be
# best to retire this interface and replace it with something more flexible.
# At the moment a best effort is made.
@@ -1072,7 +1072,7 @@ function partition_disk #<slice_size> <whole_disk_name> <total_slices>
# dirnum: the maximum number of subdirectories to use, -1 no limit
# filenum: the maximum number of files per subdirectory
# bytes: number of bytes to write
-# num_writes: numer of types to write out bytes
+# num_writes: number of types to write out bytes
# data: the data that will be written
#
# E.g.
@@ -2859,7 +2859,7 @@ function labelvtoc
#
# check if the system was installed as zfsroot or not
-# return: 0 ture, otherwise false
+# return: 0 if zfsroot, non-zero if not
#
function is_zfsroot
{
diff --git a/tests/zfs-tests/tests/functional/acl/acl_common.kshlib b/tests/zfs-tests/tests/functional/acl/acl_common.kshlib
index a81cd76ba..ba08bcb48 100644
--- a/tests/zfs-tests/tests/functional/acl/acl_common.kshlib
+++ b/tests/zfs-tests/tests/functional/acl/acl_common.kshlib
@@ -34,7 +34,7 @@
#
# Get the given file/directory access mode
#
-# $1 object -- file or directroy
+# $1 object -- file or directory
#
function get_mode #<obj>
{
@@ -49,7 +49,7 @@ function get_mode #<obj>
#
# Get the given file/directory ACL
#
-# $1 object -- file or directroy
+# $1 object -- file or directory
#
function get_acl #<obj>
{
@@ -64,7 +64,7 @@ function get_acl #<obj>
#
# Get the given file/directory ACL
#
-# $1 object -- file or directroy
+# $1 object -- file or directory
#
function get_compact_acl #<obj>
{
@@ -243,12 +243,12 @@ function usr_exec #<commands> [...]
#
# Count how many ACEs for the specified file or directory.
#
-# $1 file or directroy name
+# $1 file or directory name
#
function count_ACE #<file or dir name>
{
if [[ ! -e $1 ]]; then
- log_note "Need input file or directroy name."
+ log_note "Need input file or directory name."
return 1
fi
@@ -399,7 +399,7 @@ function rwx_node #user node acl_spec|access
#
# Get the given file/directory xattr
#
-# $1 object -- file or directroy
+# $1 object -- file or directory
#
function get_xattr #<obj>
{
diff --git a/tests/zfs-tests/tests/functional/cachefile/cachefile_004_pos.ksh b/tests/zfs-tests/tests/functional/cachefile/cachefile_004_pos.ksh
index e0b81e166..841b141e1 100755
--- a/tests/zfs-tests/tests/functional/cachefile/cachefile_004_pos.ksh
+++ b/tests/zfs-tests/tests/functional/cachefile/cachefile_004_pos.ksh
@@ -38,9 +38,9 @@
# Verify set, export and destroy when cachefile is set on pool.
#
# STRATEGY:
-# 1. Create two pools with one same cahcefile1.
+# 1. Create two pools with one same cachefile1.
# 2. Set cachefile of the two pools to another same cachefile2.
-# 3. Verify cachefile1 not exist.
+# 3. Verify cachefile1 does not exist.
# 4. Export the two pools.
# 5. Verify cachefile2 not exist.
# 6. Import the two pools and set cachefile to cachefile2.
diff --git a/tests/zfs-tests/tests/functional/casenorm/insensitive_formd_lookup.ksh b/tests/zfs-tests/tests/functional/casenorm/insensitive_formd_lookup.ksh
index d28431300..1ef9d2756 100755
--- a/tests/zfs-tests/tests/functional/casenorm/insensitive_formd_lookup.ksh
+++ b/tests/zfs-tests/tests/functional/casenorm/insensitive_formd_lookup.ksh
@@ -19,7 +19,7 @@
# DESCRIPTION:
# For the filesystem with casesensitivity=insensitive, normalization=formD,
-# check that lookup succeds using any name form.
+# check that lookup succeeds using any name form.
#
# STRATEGY:
# For each c/n name form:
diff --git a/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.list_user_props.ksh b/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.list_user_props.ksh
index 34fdbd56d..a454a2753 100755
--- a/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.list_user_props.ksh
+++ b/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.list_user_props.ksh
@@ -144,4 +144,4 @@ log_must_program $TESTPOOL - <<-EOF
return 0
EOF
-log_pass "Listing zfs user properies should work correctly."
+log_pass "Listing zfs user properties should work correctly."
diff --git a/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.terminate_by_signal.ksh b/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.terminate_by_signal.ksh
index 6f58cc1f4..74889eba8 100755
--- a/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.terminate_by_signal.ksh
+++ b/tests/zfs-tests/tests/functional/channel_program/synctask_core/tst.terminate_by_signal.ksh
@@ -90,7 +90,7 @@ snap_count=$(zfs list -t snapshot | grep $TESTPOOL | wc -l)
log_note "$snap_count snapshots created by ZCP"
if [ "$snap_count" -eq 0 ]; then
- log_fail "Channel progam failed to run."
+ log_fail "Channel program failed to run."
elif [ "$snap_count" -gt 50 ]; then
log_fail "Too many snapshots after a cancel ($snap_count)."
else
diff --git a/tests/zfs-tests/tests/functional/cli_root/zfs_clone/zfs_clone_010_pos.ksh b/tests/zfs-tests/tests/functional/cli_root/zfs_clone/zfs_clone_010_pos.ksh
index 40cabf649..62a755eae 100755
--- a/tests/zfs-tests/tests/functional/cli_root/zfs_clone/zfs_clone_010_pos.ksh
+++ b/tests/zfs-tests/tests/functional/cli_root/zfs_clone/zfs_clone_010_pos.ksh
@@ -150,7 +150,7 @@ log_note "Verify zfs clone property for multiple clones"
names=$(zfs list -rt all -o name $TESTPOOL)
log_must verify_clones 3 0
-log_note "verfify clone property for clone deletion"
+log_note "verify clone property for clone deletion"
i=1
for ds in $datasets; do
log_must zfs destroy $ds/$TESTCLONE.$i
diff --git a/tests/zfs-tests/tests/functional/cli_root/zfs_copies/zfs_copies_002_pos.ksh b/tests/zfs-tests/tests/functional/cli_root/zfs_copies/zfs_copies_002_pos.ksh
index a5a9729dc..11265cd5a 100755
--- a/tests/zfs-tests/tests/functional/cli_root/zfs_copies/zfs_copies_002_pos.ksh
+++ b/tests/zfs-tests/tests/functional/cli_root/zfs_copies/zfs_copies_002_pos.ksh
@@ -92,7 +92,7 @@ for val in 1 2 3; do
check_used $used $val
done
-log_note "Verify df(1M) can corectly display the space charged."
+log_note "Verify df(1M) can correctly display the space charged."
for val in 1 2 3; do
used=`df -F zfs -k /$TESTPOOL/fs_$val/$FILE | grep $TESTPOOL/fs_$val \
| awk '{print $3}'`
diff --git a/tests/zfs-tests/tests/functional/cli_root/zfs_create/zfs_create_011_pos.ksh b/tests/zfs-tests/tests/functional/cli_root/zfs_create/zfs_create_011_pos.ksh
index 0144b050d..982a4ea16 100755
--- a/tests/zfs-tests/tests/functional/cli_root/zfs_create/zfs_create_011_pos.ksh
+++ b/tests/zfs-tests/tests/functional/cli_root/zfs_create/zfs_create_011_pos.ksh
@@ -33,7 +33,7 @@
#
# DESCRIPTION:
-# 'zfs create -p' should work as expecteed
+# 'zfs create -p' should work as expected
#
# STRATEGY:
# 1. To create $newdataset with -p option, first make sure the upper level
diff --git a/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_001_pos.ksh b/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_001_pos.ksh
index 534c33f0a..26857d48d 100755
--- a/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_001_pos.ksh
+++ b/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_001_pos.ksh
@@ -53,7 +53,7 @@ verify_runnable "both"
# run 'zfs destroy $opt <dataset>'. 3rd, check the system status.
#
# $1 option of 'zfs destroy'
-# $2 dataset will be destroied.
+# $2 dataset will be destroyed.
#
function test_n_check
{
diff --git a/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_005_neg.ksh b/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_005_neg.ksh
index 2e4a0c3b2..1c5b2cf1c 100755
--- a/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_005_neg.ksh
+++ b/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_005_neg.ksh
@@ -145,8 +145,8 @@ if is_global_zone; then
check_dataset datasetexists $CTR $VOL
check_dataset datasetnonexists $VOLSNAP $VOLCLONE
- # Due to recusive destroy being a best-effort operation,
- # all of the non-busy datasets bellow should be gone now.
+ # Due to recursive destroy being a best-effort operation,
+ # all of the non-busy datasets below should be gone now.
check_dataset datasetnonexists $FS $FSSNAP $FSCLONE
fi
diff --git a/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_014_pos.ksh b/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_014_pos.ksh
index df7cfcf52..58c4cfb56 100755
--- a/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_014_pos.ksh
+++ b/tests/zfs-tests/tests/functional/cli_root/zfs_destroy/zfs_destroy_014_pos.ksh
@@ -24,7 +24,7 @@
#
# DESCRIPTION:
# 'zfs destroy -R <snapshot>' can destroy all the child
-# snapshots and preserves all the nested datasetss.
+# snapshots and preserves all the nested datasets.
#
# STRATEGY:
# 1. Create nested datasets in the storage pool.
@@ -57,7 +57,7 @@ for ds in $datasets; do
datasetexists $ds || log_fail "Create $ds dataset fail."
done
-# create recursive nestedd snapshot
+# create recursive nested snapshot
log_must zfs snapshot -r $TESTPOOL/$TESTFS1@snap
for ds in $datasets; do
datasetexists $ds@snap || log_fail "Create $ds@snap snapshot fail."