aboutsummaryrefslogtreecommitdiffstats
path: root/scripts
diff options
context:
space:
mode:
authorBrian Behlendorf <[email protected]>2015-07-01 15:23:09 -0700
committerBrian Behlendorf <[email protected]>2016-03-16 13:46:16 -0700
commit6bb24f4dc7b7267699e3c3a4ca1ca062fe564b9e (patch)
treeaabbb51e18eadc173599f8ed700ab98e198d6ffd /scripts
parent887d1e60ef1f20a1b13e7c7b0d208f10b13b9cbe (diff)
Add the ZFS Test Suite
Add the ZFS Test Suite and test-runner framework from illumos. This is a continuation of the work done by Turbo Fredriksson to port the ZFS Test Suite to Linux. While this work was originally conceived as a stand alone project integrating it directly with the ZoL source tree has several advantages: * Allows the ZFS Test Suite to be packaged in zfs-test package. * Facilitates easy integration with the CI testing. * Users can locally run the ZFS Test Suite to validate ZFS. This testing should ONLY be done on a dedicated test system because the ZFS Test Suite in its current form is destructive. * Allows the ZFS Test Suite to be run directly in the ZoL source tree enabled developers to iterate quickly during development. * Developers can easily add/modify tests in the framework as features are added or functionality is changed. The tests will then always be in sync with the implementation. Full documentation for how to run the ZFS Test Suite is available in the tests/README.md file. Warning: This test suite is designed to be run on a dedicated test system. It will make modifications to the system including, but not limited to, the following. * Adding new users * Adding new groups * Modifying the following /proc files: * /proc/sys/kernel/core_pattern * /proc/sys/kernel/core_uses_pid * Creating directories under / Notes: * Not all of the test cases are expected to pass and by default these test cases are disabled. The failures are primarily due to assumption made for illumos which are invalid under Linux. * When updating these test cases it should be done in as generic a way as possible so the patch can be submitted back upstream. Most existing library functions have been updated to be Linux aware, and the following functions and variables have been added. * Functions: * is_linux - Used to wrap a Linux specific section. * block_device_wait - Waits for block devices to be added to /dev/. * Variables: Linux Illumos * ZVOL_DEVDIR "/dev/zvol" "/dev/zvol/dsk" * ZVOL_RDEVDIR "/dev/zvol" "/dev/zvol/rdsk" * DEV_DSKDIR "/dev" "/dev/dsk" * DEV_RDSKDIR "/dev" "/dev/rdsk" * NEWFS_DEFAULT_FS "ext2" "ufs" * Many of the disabled test cases fail because 'zfs/zpool destroy' returns EBUSY. This is largely causes by the asynchronous nature of device handling on Linux and is expected, the impacted test cases will need to be updated to handle this. * There are several test cases which have been disabled because they can trigger a deadlock. A primary example of this is to recursively create zpools within zpools. These tests have been disabled until the root issue can be addressed. * Illumos specific utilities such as (mkfile) should be added to the tests/zfs-tests/cmd/ directory. Custom programs required by the test scripts can also be added here. * SELinux should be either is permissive mode or disabled when running the tests. The test cases should be updated to conform to a standard policy. * Redundant test functionality has been removed (zfault.sh). * Existing test scripts (zconfig.sh) should be migrated to use the framework for consistency and ease of testing. * The DISKS environment variable currently only supports loopback devices because of how the ZFS Test Suite expects partitions to be named (p1, p2, etc). Support must be added to generate the correct partition name based on the device location and name. * The ZFS Test Suite is part of the illumos code base at: https://github.com/illumos/illumos-gate/tree/master/usr/src/test Original-patch-by: Turbo Fredriksson <[email protected]> Signed-off-by: Brian Behlendorf <[email protected]> Signed-off-by: Olaf Faaland <[email protected]> Closes #6 Closes #1534
Diffstat (limited to 'scripts')
-rw-r--r--scripts/Makefile.am42
-rw-r--r--scripts/common.sh.in11
-rwxr-xr-xscripts/zfault.sh955
-rwxr-xr-xscripts/zfs-tests.sh343
4 files changed, 353 insertions, 998 deletions
diff --git a/scripts/Makefile.am b/scripts/Makefile.am
index 26f9207e4..32d8911c3 100644
--- a/scripts/Makefile.am
+++ b/scripts/Makefile.am
@@ -6,52 +6,12 @@ pkgdatadir = $(datadir)/@PACKAGE@
dist_pkgdata_SCRIPTS = \
$(top_builddir)/scripts/common.sh \
$(top_srcdir)/scripts/zconfig.sh \
- $(top_srcdir)/scripts/zfault.sh \
$(top_srcdir)/scripts/ziltest.sh \
$(top_srcdir)/scripts/zimport.sh \
$(top_srcdir)/scripts/zfs.sh \
+ $(top_srcdir)/scripts/zfs-tests.sh \
$(top_srcdir)/scripts/zpool-create.sh \
$(top_srcdir)/scripts/zpios.sh \
$(top_srcdir)/scripts/zpios-sanity.sh \
$(top_srcdir)/scripts/zpios-survey.sh \
$(top_srcdir)/scripts/smb.sh
-
-ZFS=$(top_builddir)/scripts/zfs.sh
-ZCONFIG=$(top_builddir)/scripts/zconfig.sh
-ZFAULT=$(top_builddir)/scripts/zfault.sh
-ZIMPORT=$(top_builddir)/scripts/zimport.sh
-ZTEST=$(top_builddir)/cmd/ztest/ztest
-ZPIOS_SANITY=$(top_builddir)/scripts/zpios-sanity.sh
-
-check:
- @$(ZFS) -u
- @echo
- @echo -n "===================================="
- @echo -n " ZTEST "
- @echo "===================================="
- @echo
- @$(ZFS)
- @$(ZTEST) -V
- @$(ZFS) -u
- @echo
- @echo
- @echo -n "==================================="
- @echo -n " ZCONFIG "
- @echo "==================================="
- @echo
- @$(ZCONFIG) -c
- @echo
- @echo -n "==================================="
- @echo -n " ZFAULT "
- @echo "==================================="
- @echo
- @$(ZFAULT) -c
- @echo
- @echo -n "===================================="
- @echo -n " ZPIOS "
- @echo "===================================="
- @echo
- @$(ZFS)
- @$(ZPIOS_SANITY)
- @$(ZFS) -u
- @echo
diff --git a/scripts/common.sh.in b/scripts/common.sh.in
index 2fac2a919..f6c6d93fe 100644
--- a/scripts/common.sh.in
+++ b/scripts/common.sh.in
@@ -45,10 +45,13 @@ DEVDIR=${DEVDIR:-/dev/disk/by-vdev}
ZPOOLDIR=${ZPOOLDIR:-${pkgdatadir}/zpool-config}
ZPIOSDIR=${ZPIOSDIR:-${pkgdatadir}/zpios-test}
ZPIOSPROFILEDIR=${ZPIOSPROFILEDIR:-${pkgdatadir}/zpios-profile}
+TESTSDIR=${TESTSDIR:-${pkgdatadir}/zfs-tests}
+RUNFILEDIR=${RUNFILEDIR:-${pkgdatadir}/runfiles}
ZDB=${ZDB:-${sbindir}/zdb}
ZFS=${ZFS:-${sbindir}/zfs}
ZINJECT=${ZINJECT:-${sbindir}/zinject}
+ZHACK=${ZHACK:-${sbindir}/zhack}
ZPOOL=${ZPOOL:-${sbindir}/zpool}
ZTEST=${ZTEST:-${sbindir}/ztest}
ZPIOS=${ZPIOS:-${sbindir}/zpios}
@@ -58,6 +61,9 @@ ZFS_SH=${ZFS_SH:-${pkgdatadir}/zfs.sh}
ZPOOL_CREATE_SH=${ZPOOL_CREATE_SH:-${pkgdatadir}/zpool-create.sh}
ZPIOS_SH=${ZPIOS_SH:-${pkgdatadir}/zpios.sh}
ZPIOS_SURVEY_SH=${ZPIOS_SURVEY_SH:-${pkgdatadir}/zpios-survey.sh}
+TEST_RUNNER=${TEST_RUNNER:-${pkgdatadir}/test-runner/bin/test-runner.py}
+STF_TOOLS=${STF_TOOLS:-${pkgdatadir}/test-runner}
+STF_SUITE=${STF_SUITE:-${pkgdatadir}/zfs-tests}
LDMOD=${LDMOD:-/sbin/modprobe}
LSMOD=${LSMOD:-/sbin/lsmod}
@@ -65,6 +71,7 @@ RMMOD=${RMMOD:-/sbin/rmmod}
INFOMOD=${INFOMOD:-/sbin/modinfo}
LOSETUP=${LOSETUP:-/sbin/losetup}
MDADM=${MDADM:-/sbin/mdadm}
+DMSETUP=${DMSETUP:-/sbin/dmsetup}
PARTED=${PARTED:-/sbin/parted}
BLOCKDEV=${BLOCKDEV:-/sbin/blockdev}
LSSCSI=${LSSCSI:-/usr/bin/lsscsi}
@@ -297,7 +304,7 @@ check_loop_utils() {
# the minor as long as it's less than /sys/module/loop/parameters/max_loop.
#
unused_loop_device() {
- local DEVICE=`${LOSETUP} -f`
+ local DEVICE=$(${LOSETUP} -f)
local MAX_LOOP_PATH="/sys/module/loop/parameters/max_loop"
local MAX_LOOP;
@@ -359,7 +366,7 @@ destroy_loop_devices() {
local LODEVICES="$1"
msg "Destroying ${LODEVICES}"
- ${LOSETUP} -d ${LODEVICES} || \
+ ${LOSETUP} -d ${LODEVICES} || \
die "Error $? destroying ${FILE} -> ${DEVICE} loopback"
rm -f ${FILES}
diff --git a/scripts/zfault.sh b/scripts/zfault.sh
deleted file mode 100755
index a5f1f3cb1..000000000
--- a/scripts/zfault.sh
+++ /dev/null
@@ -1,955 +0,0 @@
-#!/bin/bash
-#
-# ZPOOL fault verification test script.
-#
-# The current suite of fault tests should not be thought of an exhaustive
-# list of failure modes. Rather it is simply an starting point which trys
-# to cover the bulk the of the 'easy' and hopefully common, failure modes.
-#
-# Additional tests should be added but the current suite as new interesting
-# failures modes are observed. Additional failure modes I'd like to see
-# tests for include, but are not limited too:
-#
-# * Slow but successful IO.
-# * SCSI sense codes generated as zevents.
-# * 4k sectors
-# * noise
-# * medium error
-# * recovered error
-#
-# The current infrastructure using the 'mdadm' faulty device and the
-# 'scsi_debug' simulated scsi devices. The idea is to inject the error
-# below the zfs stack to validate all the error paths. More targeted
-# failure testing should be added using the 'zinject' command line util.
-#
-# Requires the following packages:
-# * mdadm
-# * lsscsi
-# * sg3-utils
-#
-
-basedir="$(dirname $0)"
-
-SCRIPT_COMMON=common.sh
-if [ -f "${basedir}/${SCRIPT_COMMON}" ]; then
-. "${basedir}/${SCRIPT_COMMON}"
-else
-echo "Missing helper script ${SCRIPT_COMMON}" && exit 1
-fi
-
-PROG=zfault.sh
-
-usage() {
-cat << EOF
-USAGE:
-$0 [hvcts]
-
-DESCRIPTION:
- ZPOOL fault verification tests
-
-OPTIONS:
- -h Show this message
- -v Verbose
- -c Cleanup md+lo+file devices at start
- -t <#> Run listed tests
- -s <#> Skip listed tests
-
-EOF
-}
-
-while getopts 'hvct:s:?' OPTION; do
- case $OPTION in
- h)
- usage
- exit 1
- ;;
- v)
- VERBOSE=1
- ;;
- c)
- CLEANUP=1
- ;;
- t)
- TESTS_RUN=($OPTARG)
- ;;
- s)
- TESTS_SKIP=($OPTARG)
- ;;
- ?)
- usage
- exit
- ;;
- esac
-done
-
-if [ $(id -u) != 0 ]; then
- die "Must run as root"
-fi
-
-# Initialize the test suite
-init
-
-# Perform pre-cleanup is requested
-if [ ${CLEANUP} ]; then
- ${ZFS_SH} -u
- cleanup_md_devices
- cleanup_loop_devices
- rm -f /tmp/zpool.cache.*
-fi
-
-# Check if we need to skip all md based tests.
-MD_PARTITIONABLE=0
-check_md_partitionable && MD_PARTITIONABLE=1
-if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- echo "Skipping tests 1-7 which require partitionable md devices"
-fi
-
-# Check if we need to skip all the scsi_debug tests.
-SCSI_DEBUG=0
-${INFOMOD} scsi_debug &>/dev/null && SCSI_DEBUG=1
-if [ ${SCSI_DEBUG} -eq 0 ]; then
- echo "Skipping tests 8-9 which require the scsi_debug module"
-fi
-
-if [ ${MD_PARTITIONABLE} -eq 0 ] || [ ${SCSI_DEBUG} -eq 0 ]; then
- echo
-fi
-
-printf "%40s%s\t%s\t%s\t%s\t%s\n" "" "raid0" "raid10" "raidz" "raidz2" "raidz3"
-
-pass_nonewline() {
- echo -n -e "${COLOR_GREEN}Pass${COLOR_RESET}\t"
-}
-
-skip_nonewline() {
- echo -n -e "${COLOR_BROWN}Skip${COLOR_RESET}\t"
-}
-
-nth_zpool_vdev() {
- local POOL_NAME=$1
- local DEVICE_TYPE=$2
- local DEVICE_NTH=$3
-
- ${ZPOOL} status ${POOL_NAME} | grep ${DEVICE_TYPE} ${TMP_STATUS} | \
- head -n${DEVICE_NTH} | tail -n1 | ${AWK} "{ print \$1 }"
-}
-
-vdev_status() {
- local POOL_NAME=$1
- local VDEV_NAME=$2
-
- ${ZPOOL} status ${POOL_NAME} | ${AWK} "/${VDEV_NAME}/ { print \$2 }"
-}
-
-# Required format is x.yz[KMGTP]
-expand_numeric_suffix() {
- local VALUE=$1
-
- VALUE=`echo "${VALUE/%K/*1000}"`
- VALUE=`echo "${VALUE/%M/*1000000}"`
- VALUE=`echo "${VALUE/%G/*1000000000}"`
- VALUE=`echo "${VALUE/%T/*1000000000000}"`
- VALUE=`echo "${VALUE/%P/*1000000000000000}"`
- VALUE=`echo "${VALUE}" | bc | cut -d'.' -f1`
-
- echo "${VALUE}"
-}
-
-vdev_read_errors() {
- local POOL_NAME=$1
- local VDEV_NAME=$2
- local VDEV_ERRORS=`${ZPOOL} status ${POOL_NAME} |
- ${AWK} "/${VDEV_NAME}/ { print \\$3 }"`
-
- expand_numeric_suffix ${VDEV_ERRORS}
-}
-
-vdev_write_errors() {
- local POOL_NAME=$1
- local VDEV_NAME=$2
- local VDEV_ERRORS=`${ZPOOL} status ${POOL_NAME} |
- ${AWK} "/${VDEV_NAME}/ { print \\$4 }"`
-
- expand_numeric_suffix ${VDEV_ERRORS}
-}
-
-vdev_cksum_errors() {
- local POOL_NAME=$1
- local VDEV_NAME=$2
- local VDEV_ERRORS=`${ZPOOL} status ${POOL_NAME} |
- ${AWK} "/${VDEV_NAME}/ { print \\$5 }"`
-
- expand_numeric_suffix ${VDEV_ERRORS}
-}
-
-zpool_state() {
- local POOL_NAME=$1
-
- ${ZPOOL} status ${POOL_NAME} | ${AWK} "/state/ { print \$2; exit }"
-}
-
-zpool_event() {
- local EVENT_NAME=$1
- local EVENT_KEY=$2
-
- SCRIPT1="BEGIN {RS=\"\"; FS=\"\n\"} /${EVENT_NAME}/ { print \$0; exit }"
- SCRIPT2="BEGIN {FS=\"=\"} /${EVENT_KEY}/ { print \$2; exit }"
-
- ${ZPOOL} events -vH | ${AWK} "${SCRIPT1}" | ${AWK} "${SCRIPT2}"
-}
-
-zpool_scan_errors() {
- local POOL_NAME=$1
-
- ${ZPOOL} status ${POOL_NAME} | ${AWK} "/scan: scrub/ { print \$8 }"
- ${ZPOOL} status ${POOL_NAME} | ${AWK} "/scan: resilver/ { print \$7 }"
-}
-
-pattern_create() {
- local PATTERN_BLOCK_SIZE=$1
- local PATTERN_BLOCK_COUNT=$2
- local PATTERN_NAME=`mktemp -p /tmp zpool.pattern.XXXXXXXX`
-
- echo ${PATTERN_NAME}
- dd if=/dev/urandom of=${PATTERN_NAME} bs=${PATTERN_BLOCK_SIZE} \
- count=${PATTERN_BLOCK_COUNT} &>/dev/null
- return $?
-}
-
-pattern_write() {
- local PATTERN_NAME=$1
- local PATTERN_BLOCK_SIZE=$2
- local PATTERN_BLOCK_COUNT=$3
- local DEVICE_NAME=$4
-
- dd if=${PATTERN_NAME} of=${DEVICE_NAME} bs=${PATTERN_BLOCK_SIZE} \
- count=${PATTERN_BLOCK_COUNT} oflag=direct &>/dev/null
- return $?
-}
-
-pattern_write_bg() {
- local PATTERN_NAME=$1
- local PATTERN_BLOCK_SIZE=$2
- local PATTERN_BLOCK_COUNT=$3
- local DEVICE_NAME=$4
-
- dd if=${PATTERN_NAME} of=${DEVICE_NAME} bs=${PATTERN_BLOCK_SIZE} \
- count=${PATTERN_BLOCK_COUNT} oflag=direct &>/dev/null &
- return $?
-}
-
-pattern_verify() {
- local PATTERN_NAME=$1
- local PATTERN_BLOCK_SIZE=$2
- local PATTERN_BLOCK_COUNT=$3
- local DEVICE_NAME=$4
- local DEVICE_FILE=`mktemp -p /tmp zpool.pattern.XXXXXXXX`
-
- dd if=${DEVICE_NAME} of=${DEVICE_FILE} bs=${PATTERN_BLOCK_SIZE} \
- count=${PATTERN_BLOCK_COUNT} iflag=direct &>/dev/null
- cmp -s ${PATTERN_NAME} ${DEVICE_FILE}
- RC=$?
- rm -f ${DEVICE_FILE}
-
- return ${RC}
-}
-
-pattern_remove() {
- local PATTERN_NAME=$1
-
- rm -f ${PATTERN_NAME}
- return $?
-}
-
-fault_set_md() {
- local VDEV_FAULTY=$1
- local FAULT_TYPE=$2
-
- ${MDADM} /dev/${VDEV_FAULTY} --grow --level=faulty \
- --layout=${FAULT_TYPE} >/dev/null
- return $?
-}
-
-fault_clear_md() {
- local VDEV_FAULTY=$1
-
- # Clear all failure injection.
- ${MDADM} /dev/${VDEV_FAULTY} --grow --level=faulty \
- --layout=clear >/dev/null || return $?
- ${MDADM} /dev/${VDEV_FAULTY} --grow --level=faulty \
- --layout=flush >/dev/null || return $?
- return $?
-}
-
-fault_set_sd() {
- local OPTS=$1
- local NTH=$2
-
- echo ${OPTS} >/sys/bus/pseudo/drivers/scsi_debug/opts
- echo ${NTH} >/sys/bus/pseudo/drivers/scsi_debug/every_nth
-}
-
-fault_clear_sd() {
- echo 0 >/sys/bus/pseudo/drivers/scsi_debug/every_nth
- echo 0 >/sys/bus/pseudo/drivers/scsi_debug/opts
-}
-
-test_setup() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local ZVOL_NAME=$3
- local TMP_CACHE=$4
-
- ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1
- ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c ${POOL_CONFIG} || fail 2
- ${ZFS} create -V 64M ${POOL_NAME}/${ZVOL_NAME} || fail 3
-
- # Trigger udev and re-read the partition table to ensure all of
- # this IO is out of the way before we begin injecting failures.
- udev_trigger || fail 4
- ${BLOCKDEV} --rereadpt /dev/${POOL_NAME}/${ZVOL_NAME} || fail 5
-}
-
-test_cleanup() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local ZVOL_NAME=$3
- local TMP_CACHE=$4
-
- ${ZFS} destroy ${POOL_NAME}/${ZVOL_NAME} || fail 101
- ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c ${POOL_CONFIG} -d || fail 102
- ${ZFS_SH} -u || fail 103
- rm -f ${TMP_CACHE} || fail 104
-}
-
-test_write_soft() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
-
- if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Set soft write failure for first vdev device.
- local VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md 1`
- fault_set_md ${VDEV_FAULTY} write-transient
-
- # The application must not observe an error.
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
- fault_clear_md ${VDEV_FAULTY}
-
- # Soft errors will not be logged to 'zpool status'
- local WRITE_ERRORS=`vdev_write_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${WRITE_ERRORS} -eq 0 || fail 13
-
- # Soft errors will still generate an EIO (5) event.
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 14
-
- # Verify the known pattern.
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 15
- pattern_remove ${TMP_PATTERN} || fail 16
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-# Soft write error.
-test_1() {
- test_write_soft tank lo-faulty-raid0 0
- test_write_soft tank lo-faulty-raid10 1
- test_write_soft tank lo-faulty-raidz 1
- test_write_soft tank lo-faulty-raidz2 1
- test_write_soft tank lo-faulty-raidz3 1
- echo
-}
-run_test 1 "soft write error"
-
-test_write_hard() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
-
- if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Set hard write failure for first vdev device.
- local VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md 1`
- fault_set_md ${VDEV_FAULTY} write-persistent
-
- # The application must not observe an error.
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
- fault_clear_md ${VDEV_FAULTY}
-
- local WRITE_ERRORS=`vdev_write_errors ${POOL_NAME} ${VDEV_FAULTY}`
- if [ ${POOL_REDUNDANT} -eq 1 ]; then
- # For redundant configurations hard errors will not be
- # logged to 'zpool status' but will generate EIO events.
- test ${WRITE_ERRORS} -eq 0 || fail 21
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 22
- else
- # For non-redundant configurations hard errors will be
- # logged to 'zpool status' and generate EIO events. They
- # will also trigger a scrub of the impacted sectors.
- sleep 10
- test ${WRITE_ERRORS} -gt 0 || fail 31
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 32
- test `zpool_event "zfs.resilver.start" "ena"` != "" || fail 33
- test `zpool_event "zfs.resilver.finish" "ena"` != "" || fail 34
- test `zpool_scan_errors ${POOL_NAME}` -eq 0 || fail 35
- fi
-
- # Verify the known pattern.
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 41
- pattern_remove ${TMP_PATTERN} || fail 42
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-# Hard write error.
-test_2() {
- test_write_hard tank lo-faulty-raid0 0
- test_write_hard tank lo-faulty-raid10 1
- test_write_hard tank lo-faulty-raidz 1
- test_write_hard tank lo-faulty-raidz2 1
- test_write_hard tank lo-faulty-raidz3 1
- echo
-}
-run_test 2 "hard write error"
-
-test_write_all() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
-
- if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Set all write failures for first vdev device.
- local VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md 1`
- fault_set_md ${VDEV_FAULTY} write-all
-
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- if [ ${POOL_REDUNDANT} -eq 1 ]; then
- # The application must not observe an error.
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
- else
- # The application is expected to hang in the background until
- # the faulty device is repaired and 'zpool clear' is run.
- pattern_write_bg ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 13
- sleep 10
- fi
- fault_clear_md ${VDEV_FAULTY}
-
- local WRITE_ERRORS=`vdev_write_errors ${POOL_NAME} ${VDEV_FAULTY}`
- local VDEV_STATUS=`vdev_status ${POOL_NAME} ${VDEV_FAULTY}`
- local POOL_STATE=`zpool_state ${POOL_NAME}`
- # For all configurations write errors are logged to 'zpool status',
- # and EIO events are generated. However, only a redundant config
- # will cause the vdev to be FAULTED and pool DEGRADED. In a non-
- # redundant config the IO will hang until 'zpool clear' is run.
- test ${WRITE_ERRORS} -gt 0 || fail 14
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 15
-
- if [ ${POOL_REDUNDANT} -eq 1 ]; then
- test "${VDEV_STATUS}" = "FAULTED" || fail 21
- test "${POOL_STATE}" = "DEGRADED" || fail 22
- else
- BLOCKED=`ps a | grep "${ZVOL_DEVICE}" | grep -c -v "grep"`
- ${ZPOOL} clear ${POOL_NAME} || fail 31
- test ${BLOCKED} -eq 1 || fail 32
- wait
- fi
-
- # Verify the known pattern.
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 41
- pattern_remove ${TMP_PATTERN} || fail 42
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-# All write errors.
-test_3() {
- test_write_all tank lo-faulty-raid0 0
- test_write_all tank lo-faulty-raid10 1
- test_write_all tank lo-faulty-raidz 1
- test_write_all tank lo-faulty-raidz2 1
- test_write_all tank lo-faulty-raidz3 1
- echo
-}
-run_test 3 "all write errors"
-
-test_read_soft() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
- local READ_ERRORS=0
-
- if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Create a pattern to be verified during a read error.
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
-
- # Set soft read failure for all the vdevs to ensure we hit it.
- for (( i=1; i<=4; i++ )); do
- fault_set_md `nth_zpool_vdev ${POOL_NAME} md $i` read-transient
- done
-
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 13
- pattern_remove ${TMP_PATTERN} || fail 14
-
- # Clear all failure injection and sum read errors.
- for (( i=1; i<=4; i++ )); do
- local VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md $i`
- local VDEV_ERRORS=`vdev_read_errors ${POOL_NAME} ${VDEV_FAULTY}`
- let READ_ERRORS=${READ_ERRORS}+${VDEV_ERRORS}
- fault_clear_md ${VDEV_FAULTY}
- done
-
- # Soft errors will not be logged to 'zpool status'.
- test ${READ_ERRORS} -eq 0 || fail 15
-
- # Soft errors will still generate an EIO (5) event.
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 16
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-# Soft read error.
-test_4() {
- test_read_soft tank lo-faulty-raid0 0
- test_read_soft tank lo-faulty-raid10 1
- test_read_soft tank lo-faulty-raidz 1
- test_read_soft tank lo-faulty-raidz2 1
- test_read_soft tank lo-faulty-raidz3 1
- echo
-}
-run_test 4 "soft read error"
-
-test_read_hard() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
- local READ_ERRORS=0
-
- if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Create a pattern to be verified during a read error.
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
-
- # Set hard read failure for the fourth vdev.
- local VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md 4`
- fault_set_md ${VDEV_FAULTY} read-persistent
-
- # For a redundant pool there must be no IO error, for a non-redundant
- # pool we expect permanent damage and an IO error during verify, unless
- # we get exceptionally lucky and have just damaged redundant metadata.
- if [ ${POOL_REDUNDANT} -eq 1 ]; then
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 21
- local READ_ERRORS=`vdev_read_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${READ_ERRORS} -eq 0 || fail 22
- else
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE}
- ${ZPOOL} scrub ${POOL_NAME} || fail 32
- local READ_ERRORS=`vdev_read_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${READ_ERRORS} -gt 0 || fail 33
- ${ZPOOL} status -v ${POOL_NAME} | \
- grep -A8 "Permanent errors" | \
- grep -q "${POOL_NAME}" || fail 34
- fi
- pattern_remove ${TMP_PATTERN} || fail 41
-
- # Clear all failure injection and sum read errors.
- fault_clear_md ${VDEV_FAULTY}
-
- # Hard errors will generate an EIO (5) event.
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 42
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-# Hard read error.
-test_5() {
- test_read_hard tank lo-faulty-raid0 0
- test_read_hard tank lo-faulty-raid10 1
- test_read_hard tank lo-faulty-raidz 1
- test_read_hard tank lo-faulty-raidz2 1
- test_read_hard tank lo-faulty-raidz3 1
- echo
-}
-run_test 5 "hard read error"
-
-# Fixable read error.
-test_read_fixable() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
- local READ_ERRORS=0
-
- if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Create a pattern to be verified during a read error.
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
-
- # Set hard read failure for the fourth vdev.
- local VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md 4`
- fault_set_md ${VDEV_FAULTY} read-fixable
-
- # For a redundant pool there must be no IO error, for a non-redundant
- # pool we expect permanent damage and an IO error during verify, unless
- # we get exceptionally lucky and have just damaged redundant metadata.
- if [ ${POOL_REDUNDANT} -eq 1 ]; then
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 21
- local READ_ERRORS=`vdev_read_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${READ_ERRORS} -eq 0 || fail 22
- else
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE}
- ${ZPOOL} scrub ${POOL_NAME} || fail 32
- local READ_ERRORS=`vdev_read_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${READ_ERRORS} -gt 0 || fail 33
- ${ZPOOL} status -v ${POOL_NAME} | \
- grep -A8 "Permanent errors" | \
- grep -q "${POOL_NAME}" || fail 34
- fi
- pattern_remove ${TMP_PATTERN} || fail 41
-
- # Clear all failure injection and sum read errors.
- fault_clear_md ${VDEV_FAULTY}
-
- # Hard errors will generate an EIO (5) event.
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 42
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-# Read errors fixable with a write.
-test_6() {
- test_read_fixable tank lo-faulty-raid0 0
- test_read_fixable tank lo-faulty-raid10 1
- test_read_fixable tank lo-faulty-raidz 1
- test_read_fixable tank lo-faulty-raidz2 1
- test_read_fixable tank lo-faulty-raidz3 1
- echo
-}
-run_test 6 "fixable read error"
-
-test_cksum() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local VDEV_DAMAGE="$4"
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
-
- if [ ${MD_PARTITIONABLE} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Create a pattern to be verified.
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
-
- # Verify the pattern and that no vdev has cksum errors.
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 13
- for (( i=1; i<4; i++ )); do
- VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md ${i}`
- CKSUM_ERRORS=`vdev_cksum_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${CKSUM_ERRORS} -eq 0 || fail 14
- done
-
- # Corrupt the bulk of a vdev with random garbage, we damage as many
- # vdevs as we have levels of redundancy. For example for a raidz3
- # configuration we can trash 3 vdevs and still expect correct data.
- # This improves the odds that we read one of the damaged vdevs.
- for VDEV in ${VDEV_DAMAGE}; do
- VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} md $VDEV`
- pattern_write /dev/urandom 1M 64 /dev/${VDEV_FAULTY}p1
- done
-
- # Verify the pattern is still correct. For non-redundant pools
- # expect failure and for redundant pools success due to resilvering.
- if [ ${POOL_REDUNDANT} -eq 1 ]; then
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 16
- else
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} && fail 17
- fi
-
- CKSUM_ERRORS=`vdev_cksum_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${CKSUM_ERRORS} -gt 0 || fail 18
- STATUS=`vdev_status ${POOL_NAME} ${VDEV_FAULTY}`
- test "${STATUS}" = "ONLINE" || fail 19
-
- # The checksum errors must be logged as an event.
- local CKSUM_ERRORS=`zpool_event "zfs.checksum" "zio_err"`
- test ${CKSUM_ERRORS} = "0x34" || test ${CKSUM_ERRORS} = "0x0" || fail 20
-
- # Verify permant errors for non-redundant pools, and for redundant
- # pools trigger a scrub and check that all checksums have been fixed.
- if [ ${POOL_REDUNDANT} -eq 1 ]; then
- # Scrub the checksum errors and clear the faults.
- ${ZPOOL} scrub ${POOL_NAME} || fail 21
- sleep 3
- ${ZPOOL} clear ${POOL_NAME} || fail 22
-
- # Re-verify the pattern for fixed checksums.
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 23
- CKSUM_ERRORS=`vdev_cksum_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${CKSUM_ERRORS} -eq 0 || fail 24
-
- # Re-verify the entire pool for fixed checksums.
- ${ZPOOL} scrub ${POOL_NAME} || fail 25
- CKSUM_ERRORS=`vdev_cksum_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${CKSUM_ERRORS} -eq 0 || fail 26
- else
- ${ZPOOL} status -v ${POOL_NAME} | \
- grep -A8 "Permanent errors" | \
- grep -q "${POOL_NAME}/${ZVOL_NAME}" || fail 31
- ${ZPOOL} clear ${POOL_NAME} || fail 32
- fi
- pattern_remove ${TMP_PATTERN} || fail 41
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-# Silent data corruption
-test_7() {
- test_cksum tank lo-faulty-raid0 0 "1"
- test_cksum tank lo-faulty-raid10 1 "1 3"
- test_cksum tank lo-faulty-raidz 1 "4"
- test_cksum tank lo-faulty-raidz2 1 "3 4"
- test_cksum tank lo-faulty-raidz3 1 "2 3 4"
- echo
-}
-run_test 7 "silent data corruption"
-
-# Soft write timeout at the scsi device layer.
-test_write_timeout_soft() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local POOL_NTH=$4
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
-
- if [ ${SCSI_DEBUG} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- # Set timeout(0x4) for every nth command.
- fault_set_sd 4 ${POOL_NTH}
-
- # The application must not observe an error.
- local TMP_PATTERN=`pattern_create 1M 8` || fail 11
- pattern_write ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 12
- fault_clear_sd
-
- # Intermittent write timeouts even with FAILFAST set may not cause
- # an EIO (5) event. This is because how FAILFAST is handled depends
- # a log on the low level driver and the exact nature of the failure.
- # We will however see a 'zfs.delay' event logged due to the timeout.
- VDEV_DELAY=`zpool_event "zfs.delay" "zio_delay"`
- test `printf "%d" ${VDEV_DELAY}` -ge 30000 || fail 13
-
- # Verify the known pattern.
- pattern_verify ${TMP_PATTERN} 1M 8 ${ZVOL_DEVICE} || fail 14
- pattern_remove ${TMP_PATTERN} || fail 15
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-test_8() {
- test_write_timeout_soft tank scsi_debug-raid0 0 50
- test_write_timeout_soft tank scsi_debug-raid10 1 100
- test_write_timeout_soft tank scsi_debug-raidz 1 75
- test_write_timeout_soft tank scsi_debug-raidz2 1 150
- test_write_timeout_soft tank scsi_debug-raidz3 1 300
- echo
-}
-run_test 8 "soft write timeout"
-
-# Persistent write timeout at the scsi device layer.
-test_write_timeout_hard() {
- local POOL_NAME=$1
- local POOL_CONFIG=$2
- local POOL_REDUNDANT=$3
- local POOL_NTH=$4
- local ZVOL_NAME="zvol"
- local ZVOL_DEVICE="/dev/${POOL_NAME}/${ZVOL_NAME}"
- local RESCAN=1
-
- if [ ${SCSI_DEBUG} -eq 0 ]; then
- skip_nonewline
- return
- fi
-
- local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX`
- test_setup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
-
- local TMP_PATTERN1=`pattern_create 1M 8`
- local TMP_PATTERN2=`pattern_create 1M 8`
- local TMP_PATTERN3=`pattern_create 1M 8`
-
- # Create three partitions each one gets a unique pattern. The first
- # pattern is written before the failure, the second pattern during
- # the failure, and the third pattern while the vdev is degraded.
- # All three patterns are verified while the vdev is degraded and
- # then again once it is brought back online.
- ${PARTED} -s ${ZVOL_DEVICE} mklabel gpt || fail 11
- ${PARTED} -s ${ZVOL_DEVICE} mkpart primary 1M 16M || fail 12
- ${PARTED} -s ${ZVOL_DEVICE} mkpart primary 16M 32M || fail 13
- ${PARTED} -s ${ZVOL_DEVICE} mkpart primary 32M 48M || fail 14
-
- wait_udev ${ZVOL_DEVICE}1 30
- wait_udev ${ZVOL_DEVICE}2 30
- wait_udev ${ZVOL_DEVICE}3 30
-
- # Before the failure.
- pattern_write ${TMP_PATTERN1} 1M 8 ${ZVOL_DEVICE}1 || fail 15
-
- # Get the faulty vdev name.
- local VDEV_FAULTY=`nth_zpool_vdev ${POOL_NAME} sd 1`
-
- # Set timeout(0x4) for every nth command.
- fault_set_sd 4 ${POOL_NTH}
-
- # During the failure.
- pattern_write ${TMP_PATTERN2} 1M 8 ${ZVOL_DEVICE}2 || fail 21
-
- # Expect write errors to be logged to 'zpool status'
- local WRITE_ERRORS=`vdev_write_errors ${POOL_NAME} ${VDEV_FAULTY}`
- test ${WRITE_ERRORS} -gt 0 || fail 22
-
- local VDEV_STATUS=`vdev_status ${POOL_NAME} ${VDEV_FAULTY}`
- test "${VDEV_STATUS}" = "UNAVAIL" || fail 23
-
- # Clear the error and remove it from /dev/.
- fault_clear_sd
- rm -f /dev/${VDEV_FAULTY}[0-9]
-
- # Verify the first two patterns and write out the third.
- pattern_write ${TMP_PATTERN3} 1M 8 ${ZVOL_DEVICE}3 || fail 31
- pattern_verify ${TMP_PATTERN1} 1M 8 ${ZVOL_DEVICE}1 || fail 32
- pattern_verify ${TMP_PATTERN2} 1M 8 ${ZVOL_DEVICE}2 || fail 33
- pattern_verify ${TMP_PATTERN3} 1M 8 ${ZVOL_DEVICE}3 || fail 34
-
- # Bring the device back online by rescanning for it. It must appear
- # in lsscsi and be available to dd before allowing ZFS to bring it
- # online. This is not required but provides additional sanity.
- while [ ${RESCAN} -eq 1 ]; do
- scsi_rescan
- wait_udev /dev/${VDEV_FAULTY} 30
-
- if [ `${LSSCSI} | grep -c "/dev/${VDEV_FAULTY}"` -eq 0 ]; then
- continue
- fi
-
- dd if=/dev/${VDEV_FAULTY} of=/dev/null bs=8M count=1 &>/dev/null
- if [ $? -ne 0 ]; then
- continue
- fi
-
- RESCAN=0
- done
-
- # Bring the device back online. We expect it to be automatically
- # resilvered without error and we should see minimally the zfs.io,
- # zfs.statechange (VDEV_STATE_HEALTHY (0x7)), and zfs.resilver.*
- # events posted.
- ${ZPOOL} online ${POOL_NAME} ${VDEV_FAULTY} || fail 51
- sleep 3
- test `zpool_event "zfs.io" "zio_err"` = "0x5" || fail 52
- test `zpool_event "zfs.statechange" "vdev_state"` = "0x7" || fail 53
- test `zpool_event "zfs.resilver.start" "ena"` != "" || fail 54
- test `zpool_event "zfs.resilver.finish" "ena"` != "" || fail 55
- test `zpool_scan_errors ${POOL_NAME}` -eq 0 || fail 56
-
- local VDEV_STATUS=`vdev_status ${POOL_NAME} ${VDEV_FAULTY}`
- test "${VDEV_STATUS}" = "ONLINE" || fail 57
-
- # Verify the known pattern.
- pattern_verify ${TMP_PATTERN1} 1M 8 ${ZVOL_DEVICE}1 || fail 61
- pattern_verify ${TMP_PATTERN2} 1M 8 ${ZVOL_DEVICE}2 || fail 62
- pattern_verify ${TMP_PATTERN3} 1M 8 ${ZVOL_DEVICE}3 || fail 63
- pattern_remove ${TMP_PATTERN1} || fail 64
- pattern_remove ${TMP_PATTERN2} || fail 65
- pattern_remove ${TMP_PATTERN3} || fail 66
-
- test_cleanup ${POOL_NAME} ${POOL_CONFIG} ${ZVOL_NAME} ${TMP_CACHE}
- pass_nonewline
-}
-
-test_9() {
- skip_nonewline # Skip non-redundant config
- test_write_timeout_hard tank scsi_debug-raid10 1 -50
- test_write_timeout_hard tank scsi_debug-raidz 1 -50
- test_write_timeout_hard tank scsi_debug-raidz2 1 -50
- test_write_timeout_hard tank scsi_debug-raidz3 1 -50
- echo
-}
-run_test 9 "hard write timeout"
-
-exit 0
diff --git a/scripts/zfs-tests.sh b/scripts/zfs-tests.sh
new file mode 100755
index 000000000..0c8a56c27
--- /dev/null
+++ b/scripts/zfs-tests.sh
@@ -0,0 +1,343 @@
+#!/bin/bash
+#
+# CDDL HEADER START
+#
+# The contents of this file are subject to the terms of the
+# Common Development and Distribution License, Version 1.0 only
+# (the "License"). You may not use this file except in compliance
+# with the License.
+#
+# You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
+# or http://www.opensolaris.org/os/licensing.
+# See the License for the specific language governing permissions
+# and limitations under the License.
+#
+# When distributing Covered Code, include this CDDL HEADER in each
+# file and include the License file at usr/src/OPENSOLARIS.LICENSE.
+# If applicable, add the following below this CDDL HEADER, with the
+# fields enclosed by brackets "[]" replaced with your own identifying
+# information: Portions Copyright [yyyy] [name of copyright owner]
+#
+# CDDL HEADER END
+#
+basedir="$(dirname $0)"
+
+SCRIPT_COMMON=common.sh
+if [ -f "${basedir}/${SCRIPT_COMMON}" ]; then
+. "${basedir}/${SCRIPT_COMMON}"
+else
+echo "Missing helper script ${SCRIPT_COMMON}" && exit 1
+fi
+
+. $STF_SUITE/include/default.cfg
+
+PROG=zfs-tests.sh
+SUDO=/usr/bin/sudo
+SETENFORCE=/usr/sbin/setenforce
+VERBOSE=
+QUIET=
+CLEANUP=1
+CLEANUPALL=0
+LOOPBACK=1
+FILESIZE="2G"
+RUNFILE=${RUNFILE:-"linux.run"}
+FILEDIR=${FILEDIR:-/var/tmp}
+DISKS=${DISKS:-""}
+
+#
+# Attempt to remove loopback devices and files which where created earlier
+# by this script to run the test framework. The '-k' option may be passed
+# to the script to suppress cleanup for debugging purposes.
+#
+cleanup() {
+ if [ $CLEANUP -eq 0 ]; then
+ return 0
+ fi
+
+ if [ $LOOPBACK -eq 1 ]; then
+ for TEST_LOOPBACK in ${LOOPBACKS}; do
+ LOOP_DEV=$(basename $TEST_LOOPBACK)
+ DM_DEV=$(${SUDO} ${DMSETUP} ls 2>/dev/null | \
+ grep ${LOOP_DEV} | cut -f1)
+
+ if [ -n "$DM_DEV" ]; then
+ ${SUDO} ${DMSETUP} remove ${DM_DEV} ||
+ echo "Failed to remove: ${DM_DEV}"
+ fi
+
+ if [ -n "${TEST_LOOPBACK}" ]; then
+ ${SUDO} ${LOSETUP} -d ${TEST_LOOPBACK} ||
+ echo "Failed to remove: ${TEST_LOOPBACK}"
+ fi
+ done
+ fi
+
+ for TEST_FILE in ${FILES}; do
+ rm -f ${TEST_FILE} &>/dev/null
+ done
+}
+trap cleanup EXIT
+
+#
+# Attempt to remove all testpools (testpool.XXX), unopened dm devices,
+# loopback devices, and files. This is a useful way to cleanup a previous
+# test run failure which has left the system in an unknown state. This can
+# be dangerous and should only be used in a dedicated test environment.
+#
+cleanup_all() {
+ local TEST_POOLS=$(${SUDO} ${ZPOOL} list -H -o name | grep testpool)
+ local TEST_LOOPBACKS=$(${SUDO} ${LOSETUP} -a|grep file-vdev|cut -f1 -d:)
+ local TEST_FILES=$(ls /var/tmp/file-vdev* 2>/dev/null)
+
+ msg
+ msg "--- Cleanup ---"
+ msg "Removing pool(s): $(echo ${TEST_POOLS} | tr '\n' ' ')"
+ for TEST_POOL in $TEST_POOLS; do
+ ${SUDO} ${ZPOOL} destroy ${TEST_POOL}
+ done
+
+ msg "Removing dm(s): $(${SUDO} ${DMSETUP} ls |
+ grep loop | tr '\n' ' ')"
+ ${SUDO} ${DMSETUP} remove_all
+
+ msg "Removing loopback(s): $(echo ${TEST_LOOPBACKS} | tr '\n' ' ')"
+ for TEST_LOOPBACK in $TEST_LOOPBACKS; do
+ ${SUDO} ${LOSETUP} -d ${TEST_LOOPBACK}
+ done
+
+ msg "Removing files(s): $(echo ${TEST_FILES} | tr '\n' ' ')"
+ for TEST_FILE in $TEST_FILES; do
+ ${SUDO} rm -f ${TEST_FILE}
+ done
+}
+
+#
+# Log a failure message, cleanup, and return an error.
+#
+fail() {
+ echo -e "${PROG}: $1" >&2
+ cleanup
+ exit 1
+}
+
+#
+# Takes a name as the only arguments and looks for the following variations
+# on that name. If one is found it is returned.
+#
+# $RUNFILEDIR/<name>
+# $RUNFILEDIR/<name>.run
+# <name>
+# <name>.run
+#
+find_runfile() {
+ local NAME=$1
+ local RESULT=""
+
+ if [ -f "$RUNFILEDIR/$NAME" ]; then
+ RESULT="$RUNFILEDIR/$NAME"
+ elif [ -f "$RUNFILEDIR/$NAME.run" ]; then
+ RESULT="$RUNFILEDIR/$NAME.run"
+ elif [ -f "$NAME" ]; then
+ RESULT="$NAME"
+ elif [ -f "$NAME.run" ]; then
+ RESULT="$NAME.run"
+ fi
+
+ echo "$RESULT"
+}
+
+#
+# Output a useful usage message.
+#
+usage() {
+cat << EOF
+USAGE:
+$0 [hvqxkf] [-s SIZE] [-r RUNFILE]
+
+DESCRIPTION:
+ ZFS Test Suite launch script
+
+OPTIONS:
+ -h Show this message
+ -v Verbose zfs-tests.sh output
+ -q Quiet test-runner output
+ -x Remove all testpools, dm, lo, and files (unsafe)
+ -k Disable cleanup after test failure
+ -f Use files only, disables block device tests
+ -d DIR Use DIR for files and loopback devices
+ -s SIZE Use vdevs of SIZE (default: 4G)
+ -r RUNFILE Run tests in RUNFILE (default: linux.run)
+
+EXAMPLES:
+# Run the default (linux) suite of tests and output the configuration used.
+$0 -v
+
+# Run a smaller suite of tests designed to run more quickly.
+$0 -r linux-fast
+
+# Cleanup a previous run of the test suite prior to testing, run the
+# default (linux) suite of tests and perform no cleanup on exit.
+$0 -c
+
+EOF
+}
+
+while getopts 'hvqxkfd:s:r:?' OPTION; do
+ case $OPTION in
+ h)
+ usage
+ exit 1
+ ;;
+ v)
+ VERBOSE=1
+ ;;
+ q)
+ QUIET="-q"
+ ;;
+ x)
+ CLEANUPALL=1
+ ;;
+ k)
+ CLEANUP=0
+ ;;
+ f)
+ LOOPBACK=0
+ ;;
+ d)
+ FILEDIR="$OPTARG"
+ ;;
+ s)
+ FILESIZE="$OPTARG"
+ ;;
+ r)
+ RUNFILE="$OPTARG"
+ ;;
+ ?)
+ usage
+ exit
+ ;;
+ esac
+done
+
+shift $((OPTIND-1))
+
+FILES=${FILES:-"$FILEDIR/file-vdev0 $FILEDIR/file-vdev1 $FILEDIR/file-vdev2"}
+LOOPBACKS=${LOOPBACKS:-""}
+
+#
+# Attempt to locate the runfile describing the test workload.
+#
+if [ -n "$RUNFILE" ]; then
+ SAVED_RUNFILE="$RUNFILE"
+ RUNFILE=$(find_runfile "$RUNFILE")
+ [ -z "$RUNFILE" ] && fail "Cannot find runfile: $SAVED_RUNFILE"
+fi
+
+if [ ! -r "$RUNFILE" ]; then
+ fail "Cannot read runfile: $RUNFILE"
+fi
+
+#
+# This script should not be run as root. Instead the test user, which may
+# be a normal user account, needs to be configured such that it can
+# run commands via sudo passwordlessly.
+#
+if [ $(id -u) = "0" ]; then
+ fail "This script must not be run as root."
+fi
+
+if [ $(sudo whoami) != "root" ]; then
+ fail "Passwordless sudo access required."
+fi
+
+#
+# Verify the ZFS module stack if loaded.
+#
+${SUDO} ${ZFS_SH} &>/dev/null
+
+#
+# Attempt to cleanup all previous state for a new test run.
+#
+if [ $CLEANUPALL -ne 0 ]; then
+ cleanup_all
+fi
+
+#
+# By default preserve any existing pools
+#
+if [ -z "${KEEP}" ]; then
+ KEEP=$(${SUDO} ${ZPOOL} list -H -o name)
+ if [ -z "${KEEP}" ]; then
+ KEEP="rpool"
+ fi
+fi
+
+msg
+msg "--- Configuration ---"
+msg "Runfile: $RUNFILE"
+msg "STF_TOOLS: $STF_TOOLS"
+msg "STF_SUITE: $STF_SUITE"
+
+#
+# No DISKS have been provided so a basic file or loopback based devices
+# must be created for the test suite to use.
+#
+if [ -z "${DISKS}" ]; then
+ #
+ # Create sparse files for the test suite. These may be used
+ # directory or have loopback devices layered on them.
+ #
+ for TEST_FILE in ${FILES}; do
+ [ -f "$TEST_FILE" ] && fail "Failed file exists: ${TEST_FILE}"
+ truncate -s ${FILESIZE} ${TEST_FILE} ||
+ fail "Failed creating: ${TEST_FILE} ($?)"
+ DISKS="$DISKS$TEST_FILE "
+ done
+
+ #
+ # If requested setup loopback devices backed by the sparse files.
+ #
+ if [ $LOOPBACK -eq 1 ]; then
+ DISKS=""
+ check_loop_utils
+
+ for TEST_FILE in ${FILES}; do
+ TEST_LOOPBACK=$(${SUDO} ${LOSETUP} -f)
+ ${SUDO} ${LOSETUP} ${TEST_LOOPBACK} ${TEST_FILE} ||
+ fail "Failed: ${TEST_FILE} -> ${TEST_LOOPBACK}"
+ LOOPBACKS="${LOOPBACKS}${TEST_LOOPBACK} "
+ DISKS="$DISKS$(basename $TEST_LOOPBACK) "
+ done
+ fi
+fi
+
+NUM_DISKS=$(echo ${DISKS} | $AWK '{print NF}')
+[ $NUM_DISKS -lt 3 ] && fail "Not enough disks ($NUM_DISKS/3 minimum)"
+
+#
+# Disable SELinux until the ZFS Test Suite has been updated accordingly.
+#
+if [ -x ${SETENFORCE} ]; then
+ ${SUDO} ${SETENFORCE} permissive &>/dev/null
+fi
+
+msg "FILEDIR: $FILEDIR"
+msg "FILES: $FILES"
+msg "LOOPBACKS: $LOOPBACKS"
+msg "DISKS: $DISKS"
+msg "NUM_DISKS: $NUM_DISKS"
+msg "FILESIZE: $FILESIZE"
+msg "Keep pool(s): $KEEP"
+msg ""
+
+export STF_TOOLS
+export STF_SUITE
+export DISKS
+export KEEP
+
+msg "${TEST_RUNNER} ${QUIET} -c ${RUNFILE} -i ${STF_SUITE}"
+${TEST_RUNNER} ${QUIET} -c ${RUNFILE} -i ${STF_SUITE}
+RESULT=$?
+echo
+
+exit ${RESULT}