diff options
author | Brian Behlendorf <[email protected]> | 2017-05-19 13:08:23 -0400 |
---|---|---|
committer | Brian Behlendorf <[email protected]> | 2017-05-22 12:34:42 -0400 |
commit | 4358afa0f2cd7629de9916ce69c5e7cdd85f6803 (patch) | |
tree | 7e64cf683899965e4f47173108a513310d85d971 | |
parent | 3f03fc8df3e74b3d9ee6eb3bd8bb656898ed7272 (diff) |
Retire zconfig.sh
All of the test coverage provided by this script is now handled
as part of the ZFS Test Suite. Remove it.
Signed-off-by: Brian Behlendorf <[email protected]>
Closes #6128
-rw-r--r-- | TEST | 4 | ||||
-rw-r--r-- | scripts/Makefile.am | 1 | ||||
-rwxr-xr-x | scripts/zconfig.sh | 696 |
3 files changed, 0 insertions, 701 deletions
@@ -13,10 +13,6 @@ #TEST_ZTEST_DIR="/var/tmp/" #TEST_ZTEST_OPTIONS="-V" -### zconfig -#TEST_ZCONFIG_SKIP="yes" -TEST_ZCONFIG_OPTIONS="-c -s10" - ### zimport #TEST_ZIMPORT_SKIP="yes" #TEST_ZIMPORT_DIR="/var/tmp/zimport" diff --git a/scripts/Makefile.am b/scripts/Makefile.am index c8218db11..74b8b31a5 100644 --- a/scripts/Makefile.am +++ b/scripts/Makefile.am @@ -5,7 +5,6 @@ EXTRA_DIST = dkms.mkconf dkms.postbuild kmodtool zfs2zol-patch.sed cstyle.pl pkgdatadir = $(datadir)/@PACKAGE@ dist_pkgdata_SCRIPTS = \ $(top_builddir)/scripts/common.sh \ - $(top_srcdir)/scripts/zconfig.sh \ $(top_srcdir)/scripts/zimport.sh \ $(top_srcdir)/scripts/zfs.sh \ $(top_srcdir)/scripts/zfs-tests.sh \ diff --git a/scripts/zconfig.sh b/scripts/zconfig.sh deleted file mode 100755 index e78c80d54..000000000 --- a/scripts/zconfig.sh +++ /dev/null @@ -1,696 +0,0 @@ -#!/bin/bash -# -# ZFS/ZPOOL configuration test script. - -basedir="$(dirname $0)" - -SCRIPT_COMMON=common.sh -if [ -f "${basedir}/${SCRIPT_COMMON}" ]; then -. "${basedir}/${SCRIPT_COMMON}" -else -echo "Missing helper script ${SCRIPT_COMMON}" && exit 1 -fi - -PROG=zconfig.sh - -usage() { -cat << EOF -USAGE: -$0 [hvcts] - -DESCRIPTION: - ZFS/ZPOOL configuration tests - -OPTIONS: - -h Show this message - -v Verbose - -c Cleanup lo+file devices at start - -t <#> Run listed tests - -s <#> Skip listed tests - -EOF -} - -while getopts 'hvct:s:?' OPTION; do - case $OPTION in - h) - usage - exit 1 - ;; - v) - VERBOSE=1 - ;; - c) - CLEANUP=1 - ;; - t) - TESTS_RUN=($OPTARG) - ;; - s) - TESTS_SKIP=($OPTARG) - ;; - ?) - usage - exit - ;; - esac -done - -if [ $(id -u) != 0 ]; then - die "Must run as root" -fi - -# Initialize the test suite -init - -# Disable the udev rule 90-zfs.rules to prevent the zfs module -# stack from being loaded due to the detection of a zfs device. -# This is important because this test scripts require full control -# over when and how the modules are loaded/unloaded. A trap is -# set to ensure the udev rule is correctly replaced on exit. -RULE=${udevruledir}/90-zfs.rules -if test -e ${RULE}; then - trap "mv ${RULE}.disabled ${RULE}" INT TERM EXIT - mv ${RULE} ${RULE}.disabled -fi - -# Perform pre-cleanup is requested -if [ ${CLEANUP} ]; then - ${ZFS_SH} -u - cleanup_md_devices - cleanup_loop_devices - rm -f /tmp/zpool.cache.* -fi - -# Check if we need to skip the tests that require scsi_debug and lsscsi. -SCSI_DEBUG=0 -${INFOMOD} scsi_debug &>/dev/null && SCSI_DEBUG=1 -HAVE_LSSCSI=0 -test -f ${LSSCSI} && HAVE_LSSCSI=1 -if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ]; then - echo "Skipping test 10 which requires the scsi_debug " \ - "module and the ${LSSCSI} utility" -fi - -# Validate persistent zpool.cache configuration. -test_1() { - local POOL_NAME=test1 - local TMP_FILE1=`mktemp` - local TMP_FILE2=`mktemp` - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create a pool save its status for comparison. - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3 - - # Unload/load the module stack and verify the pool persists. - ${ZFS_SH} -u || fail 4 - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 5 - ${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 5 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6 - cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7 - - # Cleanup the test pool and temporary files - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 8 - rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 9 - ${ZFS_SH} -u || fail 10 - - pass -} -run_test 1 "persistent zpool.cache" - -# Validate ZFS disk scanning and import w/out zpool.cache configuration. -test_2() { - local POOL_NAME=test2 - local TMP_FILE1=`mktemp` - local TMP_FILE2=`mktemp` - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create a pool save its status for comparison. - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3 - - # Unload the module stack, remove the cache file, load the module - # stack and attempt to probe the disks to import the pool. As - # a cross check verify the old pool state against the imported. - ${ZFS_SH} -u || fail 4 - rm -f ${TMP_CACHE} || fail 5 - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 6 - ${ZPOOL} import -d /dev ${POOL_NAME} || fail 8 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 9 - cmp ${TMP_FILE1} ${TMP_FILE2} || fail 10 - - # Cleanup the test pool and temporary files - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 11 - rm -f ${TMP_FILE1} ${TMP_FILE2} || fail 12 - ${ZFS_SH} -u || fail 13 - - pass -} -run_test 2 "scan disks for pools to import" - -zconfig_zvol_device_stat() { - local EXPECT=$1 - local POOL_NAME=/dev/zvol/$2 - local ZVOL_NAME=/dev/zvol/$3 - local SNAP_NAME=/dev/zvol/$4 - local CLONE_NAME=/dev/zvol/$5 - local COUNT=0 - - # Briefly delay for udev - udev_trigger - - # Pool exists - stat ${POOL_NAME} &>/dev/null && let COUNT=$COUNT+1 - - # Volume and partitions - stat ${ZVOL_NAME} &>/dev/null && let COUNT=$COUNT+1 - stat ${ZVOL_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1 - stat ${ZVOL_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1 - - # Snapshot with partitions - stat ${SNAP_NAME} &>/dev/null && let COUNT=$COUNT+1 - stat ${SNAP_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1 - stat ${SNAP_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1 - - # Clone with partitions - stat ${CLONE_NAME} &>/dev/null && let COUNT=$COUNT+1 - stat ${CLONE_NAME}-part1 &>/dev/null && let COUNT=$COUNT+1 - stat ${CLONE_NAME}-part2 &>/dev/null && let COUNT=$COUNT+1 - - if [ $EXPECT -ne $COUNT ]; then - return 1 - fi - - return 0 -} - -# zpool import/export device check -# (1 volume, 2 partitions, 1 snapshot, 1 clone) -test_3() { - local POOL_NAME=tank - local ZVOL_NAME=volume - local SNAP_NAME=snap - local CLONE_NAME=clone - local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME} - local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME} - local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME} - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create a pool, volume, partition, snapshot, and clone. - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2 - ${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3 - ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3 - label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4 - partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4 - partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4 - ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5 - ${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6 - - # Verify the devices were created - zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7 - - # Export the pool - ${ZPOOL} export ${POOL_NAME} || fail 8 - - # verify the devices were removed - zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9 - - # Import the pool, wait 1 second for udev - ${ZPOOL} import ${POOL_NAME} || fail 10 - - # Verify the devices were created - zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11 - - # Toggle the snapdev and observe snapshot device links toggled - ${ZFS} set snapdev=hidden ${FULL_ZVOL_NAME} || fail 12 - - zconfig_zvol_device_stat 7 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - "invalid" ${FULL_CLONE_NAME} || fail 13 - - ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 14 - - zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 15 - - # Destroy the pool and consequently the devices - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16 - - # verify the devices were removed - zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 17 - - ${ZFS_SH} -u || fail 18 - rm -f ${TMP_CACHE} || fail 19 - - pass -} -run_test 3 "zpool import/export device" - -# zpool insmod/rmmod device check (1 volume, 1 snapshot, 1 clone) -test_4() { - POOL_NAME=tank - ZVOL_NAME=volume - SNAP_NAME=snap - CLONE_NAME=clone - FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME} - FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME} - FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME} - TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create a pool, volume, snapshot, and clone - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2 - ${ZFS} create -V 100M ${FULL_ZVOL_NAME} || fail 3 - ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3 - label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4 - partition /dev/zvol/${FULL_ZVOL_NAME} primary 1% 50% || fail 4 - partition /dev/zvol/${FULL_ZVOL_NAME} primary 51% -1 || fail 4 - ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 5 - ${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 6 - - # Verify the devices were created - zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 7 - - # Unload the modules - ${ZFS_SH} -u || fail 8 - - # Verify the devices were removed - zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 9 - - # Load the modules, list the pools to ensure they are opened - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 10 - ${ZPOOL} import -c ${TMP_CACHE} ${POOL_NAME} || fail 10 - ${ZPOOL} list &>/dev/null - - # Verify the devices were created - zconfig_zvol_device_stat 10 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 11 - - # Destroy the pool and consequently the devices - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 12 - - # Verify the devices were removed - zconfig_zvol_device_stat 0 ${POOL_NAME} ${FULL_ZVOL_NAME} \ - ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 13 - - ${ZFS_SH} -u || fail 14 - rm -f ${TMP_CACHE} || fail 15 - - pass -} -run_test 4 "zpool insmod/rmmod device" - -# ZVOL volume sanity check -test_5() { - local POOL_NAME=tank - local ZVOL_NAME=fish - local FULL_NAME=${POOL_NAME}/${ZVOL_NAME} - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create a pool and volume. - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2 - ${ZFS} create -V 800M ${FULL_NAME} || fail 3 - label /dev/zvol/${FULL_NAME} msdos || fail 4 - partition /dev/zvol/${FULL_NAME} primary 1 -1 || fail 4 - format /dev/zvol/${FULL_NAME}-part1 ext2 || fail 5 - - # Mount the ext2 filesystem and copy some data to it. - mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6 - mount /dev/zvol/${FULL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 || fail 7 - cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 8 - sync - - # Verify the copied files match the original files. - diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \ - &>/dev/null || fail 9 - - # Remove the files, umount, destroy the volume and pool. - rm -Rf /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} || fail 10 - umount /tmp/${ZVOL_NAME}-part1 || fail 11 - rmdir /tmp/${ZVOL_NAME}-part1 || fail 12 - - ${ZFS} destroy ${FULL_NAME} || fail 13 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 14 - ${ZFS_SH} -u || fail 15 - rm -f ${TMP_CACHE} || fail 16 - - pass -} -run_test 5 "zvol+ext2 volume" - -# ZVOL snapshot sanity check -test_6() { - local POOL_NAME=tank - local ZVOL_NAME=fish - local SNAP_NAME=pristine - local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME} - local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME} - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create a pool and volume. - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raid0 || fail 2 - ${ZFS} create -s -V 800M ${FULL_ZVOL_NAME} || fail 3 - ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3 - label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4 - partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4 - format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5 - - # Mount the ext2 filesystem and copy some data to it. - mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 6 - mount /dev/zvol/${FULL_ZVOL_NAME}-part1 /tmp/${ZVOL_NAME}-part1 \ - || fail 7 - - # Snapshot the pristine ext2 filesystem and mount it read-only. - ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 8 - wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 8 - mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9 - mount /dev/zvol/${FULL_SNAP_NAME}-part1 /tmp/${SNAP_NAME}-part1 \ - &>/dev/null || fail 10 - - # Copy to original volume - cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11 - sync - - # Verify the copied files match the original files, - # and the copied files do NOT appear in the snapshot. - diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \ - &>/dev/null || fail 12 - diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \ - &>/dev/null && fail 13 - - # umount, destroy the snapshot, volume, and pool. - umount /tmp/${SNAP_NAME}-part1 || fail 14 - rmdir /tmp/${SNAP_NAME}-part1 || fail 15 - ${ZFS} destroy ${FULL_SNAP_NAME} || fail 16 - - umount /tmp/${ZVOL_NAME}-part1 || fail 17 - rmdir /tmp/${ZVOL_NAME}-part1 || fail 18 - ${ZFS} destroy ${FULL_ZVOL_NAME} || fail 19 - - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 20 - ${ZFS_SH} -u || fail 21 - rm -f ${TMP_CACHE} || fail 22 - - pass -} -run_test 6 "zvol+ext2 snapshot" - -# ZVOL clone sanity check -test_7() { - local POOL_NAME=tank - local ZVOL_NAME=fish - local SNAP_NAME=pristine - local CLONE_NAME=clone - local FULL_ZVOL_NAME=${POOL_NAME}/${ZVOL_NAME} - local FULL_SNAP_NAME=${POOL_NAME}/${ZVOL_NAME}@${SNAP_NAME} - local FULL_CLONE_NAME=${POOL_NAME}/${CLONE_NAME} - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create a pool and volume. - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2 - ${ZFS} create -V 300M ${FULL_ZVOL_NAME} || fail 3 - ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME} || fail 3 - label /dev/zvol/${FULL_ZVOL_NAME} msdos || fail 4 - partition /dev/zvol/${FULL_ZVOL_NAME} primary 1 -1 || fail 4 - format /dev/zvol/${FULL_ZVOL_NAME}-part1 ext2 || fail 5 - - # Snapshot the pristine ext2 filesystem. - ${ZFS} snapshot ${FULL_SNAP_NAME} || fail 6 - wait_udev /dev/zvol/${FULL_SNAP_NAME}-part1 30 || fail 7 - - # Mount the ext2 filesystem so some data can be copied to it. - mkdir -p /tmp/${ZVOL_NAME}-part1 || fail 7 - mount /dev/zvol/${FULL_ZVOL_NAME}-part1 \ - /tmp/${ZVOL_NAME}-part1 || fail 8 - - # Mount the pristine ext2 snapshot. - mkdir -p /tmp/${SNAP_NAME}-part1 || fail 9 - mount /dev/zvol/${FULL_SNAP_NAME}-part1 \ - /tmp/${SNAP_NAME}-part1 &>/dev/null || fail 10 - - # Copy to original volume. - cp -RL ${SRC_DIR} /tmp/${ZVOL_NAME}-part1 || fail 11 - sync - - # Verify the copied files match the original files, - # and the copied files do NOT appear in the snapshot. - diff -ur ${SRC_DIR} /tmp/${ZVOL_NAME}-part1/${SRC_DIR##*/} \ - &>/dev/null || fail 12 - diff -ur ${SRC_DIR} /tmp/${SNAP_NAME}-part1/${SRC_DIR##*/} \ - &>/dev/null && fail 13 - - # Clone from the original pristine snapshot - ${ZFS} clone ${FULL_SNAP_NAME} ${FULL_CLONE_NAME} || fail 14 - wait_udev /dev/zvol/${FULL_CLONE_NAME}-part1 30 || fail 14 - mkdir -p /tmp/${CLONE_NAME}-part1 || fail 15 - mount /dev/zvol/${FULL_CLONE_NAME}-part1 \ - /tmp/${CLONE_NAME}-part1 || fail 16 - - # Verify the clone matches the pristine snapshot, - # and the files copied to the original volume are NOT there. - diff -ur /tmp/${SNAP_NAME}-part1 /tmp/${CLONE_NAME}-part1 \ - &>/dev/null || fail 17 - diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \ - &>/dev/null && fail 18 - - # Copy to cloned volume. - cp -RL ${SRC_DIR} /tmp/${CLONE_NAME}-part1 || fail 19 - sync - - # Verify the clone matches the modified original volume. - diff -ur /tmp/${ZVOL_NAME}-part1 /tmp/${CLONE_NAME}-part1 \ - &>/dev/null || fail 20 - - # umount, destroy the snapshot, volume, and pool. - umount /tmp/${CLONE_NAME}-part1 || fail 21 - rmdir /tmp/${CLONE_NAME}-part1 || fail 22 - ${ZFS} destroy ${FULL_CLONE_NAME} || fail 23 - - umount /tmp/${SNAP_NAME}-part1 || fail 24 - rmdir /tmp/${SNAP_NAME}-part1 || fail 25 - ${ZFS} destroy ${FULL_SNAP_NAME} || fail 26 - - umount /tmp/${ZVOL_NAME}-part1 || fail 27 - rmdir /tmp/${ZVOL_NAME}-part1 || fail 28 - ${ZFS} destroy ${FULL_ZVOL_NAME} || fail 29 - - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 30 - ${ZFS_SH} -u || fail 31 - rm -f ${TMP_CACHE} || fail 32 - - pass -} -run_test 7 "zvol+ext2 clone" - -# Send/Receive sanity check -test_8() { - local POOL_NAME1=tank1 - local POOL_NAME2=tank2 - local ZVOL_NAME=fish - local SNAP_NAME=snap - local FULL_ZVOL_NAME1=${POOL_NAME1}/${ZVOL_NAME} - local FULL_ZVOL_NAME2=${POOL_NAME2}/${ZVOL_NAME} - local FULL_SNAP_NAME1=${POOL_NAME1}/${ZVOL_NAME}@${SNAP_NAME} - local FULL_SNAP_NAME2=${POOL_NAME2}/${ZVOL_NAME}@${SNAP_NAME} - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - - # Create two pools and a volume - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 || fail 2 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 || fail 2 - ${ZFS} create -V 300M ${FULL_ZVOL_NAME1} || fail 3 - ${ZFS} set snapdev=visible ${FULL_ZVOL_NAME1} || fail 3 - label /dev/zvol/${FULL_ZVOL_NAME1} msdos || fail 4 - partition /dev/zvol/${FULL_ZVOL_NAME1} primary 1 -1 || fail 4 - format /dev/zvol/${FULL_ZVOL_NAME1}-part1 ext2 || fail 5 - - # Mount the ext2 filesystem and copy some data to it. - mkdir -p /tmp/${FULL_ZVOL_NAME1}-part1 || fail 6 - mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \ - /tmp/${FULL_ZVOL_NAME1}-part1 || fail 7 - cp -RL ${SRC_DIR} /tmp/${FULL_ZVOL_NAME1}-part1 || fail 8 - - # Unmount, snapshot, mount the ext2 filesystem so it may be sent. - # We only unmount to ensure the ext2 filesystem is clean. - umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 9 - ${ZFS} snapshot ${FULL_SNAP_NAME1} || fail 10 - wait_udev /dev/zvol/${FULL_SNAP_NAME1} 30 || fail 10 - mount /dev/zvol/${FULL_ZVOL_NAME1}-part1 \ - /tmp/${FULL_ZVOL_NAME1}-part1 || 11 - - # Send/receive the snapshot from POOL_NAME1 to POOL_NAME2 - (${ZFS} send ${FULL_SNAP_NAME1} | \ - ${ZFS} receive ${FULL_ZVOL_NAME2}) || fail 12 - wait_udev /dev/zvol/${FULL_ZVOL_NAME2}-part1 30 || fail 12 - - # Mount the sent ext2 filesystem. - mkdir -p /tmp/${FULL_ZVOL_NAME2}-part1 || fail 13 - mount /dev/zvol/${FULL_ZVOL_NAME2}-part1 \ - /tmp/${FULL_ZVOL_NAME2}-part1 || fail 14 - - # Verify the contents of the volumes match - diff -ur /tmp/${FULL_ZVOL_NAME1}-part1 /tmp/${FULL_ZVOL_NAME2}-part1 \ - &>/dev/null || fail 15 - - # Umount, destroy the volume and pool. - umount /tmp/${FULL_ZVOL_NAME1}-part1 || fail 16 - umount /tmp/${FULL_ZVOL_NAME2}-part1 || fail 17 - rmdir /tmp/${FULL_ZVOL_NAME1}-part1 || fail 18 - rmdir /tmp/${FULL_ZVOL_NAME2}-part1 || fail 19 - rmdir /tmp/${POOL_NAME1} || fail 20 - rmdir /tmp/${POOL_NAME2} || fail 21 - - ${ZFS} destroy ${FULL_SNAP_NAME1} || fail 22 - ${ZFS} destroy ${FULL_SNAP_NAME2} || fail 23 - ${ZFS} destroy ${FULL_ZVOL_NAME1} || fail 24 - ${ZFS} destroy ${FULL_ZVOL_NAME2} || fail 25 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME1} -c lo-raidz2 -d || fail 26 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME2} -c lo-raidz2 -d || fail 27 - ${ZFS_SH} -u || fail 28 - rm -f ${TMP_CACHE} || fail 29 - - pass -} -run_test 8 "zfs send/receive" - -# zpool event sanity check -test_9() { - local POOL_NAME=tank - local ZVOL_NAME=fish - local FULL_NAME=${POOL_NAME}/${ZVOL_NAME} - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - local TMP_EVENTS=`mktemp -p /tmp zpool.events.XXXXXXXX` - - # Create a pool and volume. - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2 - ${ZFS} create -V 300M ${FULL_NAME} || fail 3 - udev_trigger - - # Dump the events, there should be a pool create event - ${ZPOOL} events >${TMP_EVENTS} || fail 4 - MATCHES=`grep -c sysevent\.fs\.zfs\.pool_create ${TMP_EVENTS}` - [ $MATCHES -eq 1 ] || fail 5 - - # Clear the events and ensure there are none. - ${ZPOOL} events -c >/dev/null || fail 6 - ${ZPOOL} events >${TMP_EVENTS} || fail 7 - EVENTS=`wc -l ${TMP_EVENTS} | cut -f1 -d' '` - [ $EVENTS -gt 1 ] && fail 8 - - ${ZFS} destroy ${FULL_NAME} || fail 9 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 10 - ${ZFS_SH} -u || fail 11 - rm -f ${TMP_CACHE} || fail 12 - rm -f ${TMP_EVENTS} || fail 13 - - pass -} -run_test 9 "zpool events" - -zconfig_add_vdev() { - local POOL_NAME=$1 - local TYPE=$2 - local DEVICE=$3 - local TMP_FILE1=`mktemp` - local TMP_FILE2=`mktemp` - local TMP_FILE3=`mktemp` - - BASE_DEVICE=`basename ${DEVICE}` - - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} - ${ZPOOL} add -f ${POOL_NAME} ${TYPE} ${DEVICE} 2>/dev/null || return 1 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} - diff ${TMP_FILE1} ${TMP_FILE2} > ${TMP_FILE3} - - [ `wc -l ${TMP_FILE3}|${AWK} '{print $1}'` -eq 3 ] || return 1 - - PARENT_VDEV=`tail -2 ${TMP_FILE3} | head -1 | ${AWK} '{print $NF}'` - case $TYPE in - cache) - [ "${PARENT_VDEV}" = "${TYPE}" ] || return 1 - ;; - log) - [ "${PARENT_VDEV}" = "logs" ] || return 1 - ;; - esac - - if ! tail -1 ${TMP_FILE3} | - egrep -q "^>[[:space:]]+${BASE_DEVICE}[[:space:]]+ONLINE" ; then - return 1 - fi - rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_FILE3} - - return 0 -} - -# zpool add and remove sanity check -test_10() { - local POOL_NAME=tank - local TMP_CACHE=`mktemp -p /tmp zpool.cache.XXXXXXXX` - local TMP_FILE1=`mktemp` - local TMP_FILE2=`mktemp` - - if [ ${SCSI_DEBUG} -eq 0 ] || [ ${HAVE_LSSCSI} -eq 0 ] ; then - skip - return - fi - - test `${LSMOD} | grep -c scsi_debug` -gt 0 && \ - (${RMMOD} scsi_debug || exit 1) - - /sbin/modprobe scsi_debug dev_size_mb=128 || - die "Error $? creating scsi_debug device" - udev_trigger - - SDDEVICE=`${LSSCSI}|${AWK} '/scsi_debug/ { print $6; exit }'` - BASE_SDDEVICE=`basename $SDDEVICE` - - # Create a pool - ${ZFS_SH} zfs="spa_config_path=${TMP_CACHE}" || fail 1 - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 || fail 2 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE1} || fail 3 - - # Add and remove a cache vdev by full path - zconfig_add_vdev ${POOL_NAME} cache ${SDDEVICE} || fail 4 - ${ZPOOL} remove ${POOL_NAME} ${SDDEVICE} || fail 5 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 6 - cmp ${TMP_FILE1} ${TMP_FILE2} || fail 7 - sleep 1 - - # Add and remove a cache vdev by shorthand path - zconfig_add_vdev ${POOL_NAME} cache ${BASE_SDDEVICE} || fail 8 - ${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 9 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 10 - cmp ${TMP_FILE1} ${TMP_FILE2} || fail 11 - sleep 1 - - # Add and remove a log vdev - zconfig_add_vdev ${POOL_NAME} log ${BASE_SDDEVICE} || fail 12 - ${ZPOOL} remove ${POOL_NAME} ${BASE_SDDEVICE} || fail 13 - ${ZPOOL} status ${POOL_NAME} >${TMP_FILE2} || fail 14 - cmp ${TMP_FILE1} ${TMP_FILE2} || fail 15 - - ${ZPOOL_CREATE_SH} -p ${POOL_NAME} -c lo-raidz2 -d || fail 16 - ${ZFS_SH} -u || fail 17 - ${RMMOD} scsi_debug || fail 18 - - rm -f ${TMP_FILE1} ${TMP_FILE2} ${TMP_CACHE} || fail 19 - - pass -} -run_test 10 "zpool add/remove vdev" - -exit 0 |