This is to reproduce a raid6 reconstruction bug after two drives getting offline and online via hotplug. Signed-off-by: James Alandt <James.Alandt@xxxxxxx> Signed-off-by: Liu Bo <bo.li.liu@xxxxxxxxxx> --- tests/btrfs/152 | 121 ++++++++++++++++++++++++++++++++++++++++++++++++++++++ tests/btrfs/group | 1 + 2 files changed, 122 insertions(+) create mode 100755 tests/btrfs/152 diff --git a/tests/btrfs/152 b/tests/btrfs/152 new file mode 100755 index 0000000..2b3d98c --- /dev/null +++ b/tests/btrfs/152 @@ -0,0 +1,121 @@ +#! /bin/bash +# FS QA Test 152 +# +# This test is to reproduce a bug of btrfs raid6 reconstruction. +# +#----------------------------------------------------------------------- +# Copyright (c) 2017 Oracle. All Rights Reserved. +# Author: Liu Bo <bo.li.liu@xxxxxxxxxx> +# +# This program is free software; you can redistribute it and/or +# modify it under the terms of the GNU General Public License as +# published by the Free Software Foundation. +# +# This program is distributed in the hope that it would be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU General Public License for more details. +# +# You should have received a copy of the GNU General Public License +# along with this program; if not, write the Free Software Foundation, +# Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA +#----------------------------------------------------------------------- +# + +seq=`basename $0` +seqres=$RESULT_DIR/$seq +echo "QA output created by $seq" + +here=`pwd` +tmp=/tmp/$$ +status=1 # failure is the default! +trap "_cleanup; exit \$status" 0 1 2 3 15 + +_cleanup() +{ + cd / + rm -f $tmp.* + if [ $removed == 1 ]; then + _scratch_umount + _devmgt_add ${disk1_remove} + _devmgt_add ${disk2_remove} + fi +} + +# get standard environment, filters and checks +. ./common/rc +. ./common/filter + +# remove previous $seqres.full before test +rm -f $seqres.full + +# real QA test starts here + +# Modify as appropriate. +_supported_fs btrfs +_supported_os Linux +_require_scratch +_require_scratch_nocheck +_require_scratch_dev_pool 5 +_require_deletable_scratch_dev_pool + +# umount test so that we can unload btrfs module +_test_unmount +_require_btrfs_loadable + +_scratch_dev_pool_get 5 + +devs=( $SCRATCH_DEV_POOL ) +n=${#devs[@]} + +mkfs_opts="-draid6 -mraid6 -n64K -b 1G" +_scratch_pool_mkfs $mkfs_opts >> $seqres.full 2>&1 + +_scratch_mount -o nospace_cache + +# 1st 4k write +$XFS_IO_PROG -f -c "pwrite -W 0 4K" $SCRATCH_MNT/f1 | _filter_xfs_io +sync + +# hotplug disk1 +disk1=${devs[$((n-1))]} +disk1_short=`_short_dev ${disk1}` +disk1_remove=`readlink -f /sys/block/${disk1_short} | rev | cut -d "/" -f 3 | rev` +_devmgt_remove ${disk1_remove} $disk1_short +removed=1 + +# 2nd 4K write +$XFS_IO_PROG -f -c "pwrite -W 0 4K" $SCRATCH_MNT/f2 | _filter_xfs_io +sync + +# hotplug disk2 +disk2=${devs[$((n-2))]} +disk2_short=`_short_dev ${disk2}` +disk2_remove=`readlink -f /sys/block/${disk2_short} | rev | cut -d "/" -f 3 | rev` +_devmgt_remove ${disk2_remove} $disk2_short +removed=1 + +# 3rd 4K write +$XFS_IO_PROG -f -c "pwrite -W 0 4K" $SCRATCH_MNT/f3 | _filter_xfs_io +sync + +_scratch_unmount + +# bring disk1 & disk2 online +_devmgt_add ${disk1_remove} +_devmgt_add ${disk2_remove} +removed=0 + +# reload the module to simulate 'reboot' +_reload_btrfs_ko + +# have every disk available for btrfs +btrfs dev scan + +_scratch_mount -onospace_cache,degraded || _fail "mount -onospace_cache,degraded failed" + +_scratch_dev_pool_put + +# success, all done +status=0 +exit diff --git a/tests/btrfs/group b/tests/btrfs/group index a7ff7b0..1dbb66f 100644 --- a/tests/btrfs/group +++ b/tests/btrfs/group @@ -154,3 +154,4 @@ 149 auto quick send compress 150 auto quick dangerous 151 auto quick +152 auto quick -- 2.5.0 -- To unsubscribe from this list: send the line "unsubscribe fstests" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html