On Thu, Oct 31, 2024 at 03:41:33AM +0800, Zorro Lang wrote: > On Tue, Oct 29, 2024 at 01:21:34PM -0400, Brian Foster wrote: > > fstests includes decent functional tests for online growfs and > > shrink, and decent stress tests for crash and log recovery, but no > > combination of the two. This test combines bits from a typical > > growfs stress test like xfs/104 with crash recovery cycles from a > > test like generic/388. As a result, this reproduces at least a > > couple recently fixed issues related to log recovery of online > > growfs operations. > > > > Signed-off-by: Brian Foster <bfoster@xxxxxxxxxx> > > --- > > tests/xfs/609 | 81 +++++++++++++++++++++++++++++++++++++++++++++++ > > tests/xfs/609.out | 2 ++ > > 2 files changed, 83 insertions(+) > > create mode 100755 tests/xfs/609 > > create mode 100644 tests/xfs/609.out > > > > diff --git a/tests/xfs/609 b/tests/xfs/609 > > new file mode 100755 > > index 00000000..4df966f7 > > --- /dev/null > > +++ b/tests/xfs/609 > > @@ -0,0 +1,81 @@ > > +#! /bin/bash > > +# SPDX-License-Identifier: GPL-2.0 > > +# Copyright (c) 2024 Red Hat, Inc. All Rights Reserved. > > +# > > +# FS QA Test No. 609 > > +# > > +# Test XFS online growfs log recovery. > > +# > > +. ./common/preamble > > +_begin_fstest auto growfs stress shutdown log recoveryloop > > + > > +# Import common functions. > > +. ./common/filter > > + > > +_stress_scratch() > > +{ > > + procs=4 > > + nops=999999 > > + # -w ensures that the only ops are ones which cause write I/O > > + FSSTRESS_ARGS=`_scale_fsstress_args -d $SCRATCH_MNT -w -p $procs \ > > + -n $nops $FSSTRESS_AVOID` > > + $FSSTRESS_PROG $FSSTRESS_ARGS >> $seqres.full 2>&1 & > > +} > > + > > +_require_scratch > > +_require_command "$XFS_GROWFS_PROG" xfs_growfs > > +_require_command "$KILLALL_PROG" killall > > + > > +_cleanup() > > +{ > > + $KILLALL_ALL fsstress > /dev/null 2>&1 > > + wait > > + cd / > > + rm -f $tmp.* > > +} > > + > > +_scratch_mkfs_xfs | _filter_mkfs >$seqres.full 2>$tmp.mkfs > > +. $tmp.mkfs # extract blocksize and data size for scratch device > > + > > +endsize=`expr 550 \* 1048576` # stop after growing this big > > +[ `expr $endsize / $dbsize` -lt $dblocks ] || _notrun "Scratch device too small" > > + > > +nags=4 > > +size=`expr 125 \* 1048576` # 120 megabytes initially > > +sizeb=`expr $size / $dbsize` # in data blocks > > +logblks=$(_scratch_find_xfs_min_logblocks -dsize=${size} -dagcount=${nags}) > > + > > +_scratch_mkfs_xfs -lsize=${logblks}b -dsize=${size} -dagcount=${nags} \ > > + >> $seqres.full || _fail "mkfs failed" > > > This test fails on my testing machine, as [1], due to above mkfs.xfs print > a warning: > > "mkfs.xfs: small data volume, ignoring data volume stripe unit 128 and stripe width 256" > > My test device is scripted, if without the specific mkfs options, it got: > # mkfs.xfs -f $SCRATCH_DEV > meta-data=/dev/sda6 isize=512 agcount=25, agsize=1064176 blks > = sectsz=512 attr=2, projid32bit=1 > = crc=1 finobt=1, sparse=1, rmapbt=1 > = reflink=1 bigtime=1 inobtcount=1 nrext64=1 > = exchange=0 > data = bsize=4096 blocks=26604400, imaxpct=25 > = sunit=16 swidth=32 blks > naming =version 2 bsize=4096 ascii-ci=0, ftype=1, parent=0 > log =internal log bsize=4096 blocks=179552, version=2 > = sectsz=512 sunit=16 blks, lazy-count=1 > realtime =none extsz=4096 blocks=0, rtextents=0 > > But if with the specific mkfs options, it got: > > # /usr/sbin/mkfs.xfs -f -lsize=3075b -dsize=131072000 -dagcount=4 $SCRATCH_DEV > mkfs.xfs: small data volume, ignoring data volume stripe unit 128 and stripe width 256 > meta-data=/dev/sda6 isize=512 agcount=4, agsize=8000 blks > = sectsz=512 attr=2, projid32bit=1 > = crc=1 finobt=1, sparse=1, rmapbt=1 > = reflink=1 bigtime=1 inobtcount=1 nrext64=1 > = exchange=0 > data = bsize=4096 blocks=32000, imaxpct=25 > = sunit=0 swidth=0 blks > naming =version 2 bsize=4096 ascii-ci=0, ftype=1, parent=0 > log =internal log bsize=4096 blocks=3075, version=2 > = sectsz=512 sunit=0 blks, lazy-count=1 > realtime =none extsz=4096 blocks=0, rtextents=0 > > Hi Brian, if you think "ignoreing volume stripe" doesn't affect the test, we can > filter out the stderr with "2>&1". I can help to change that when I merge. > Hmm.. I don't think it should affect things. We could probably make the scratch fs a bit bigger, but the idea is to leave enough room so it can be grown a number of times. Any idea if using a particular min size fs makes that warning go away? Either way I don't think the custom stripe unit/width should make much of a difference for a grow vs. log recovery test, so I'm fine with filtering that out if that's easiest. Brian > Others looks good to me, with above confirmation: > > Reviewed-by: Zorro Lang <zlang@xxxxxxxxxx> > > Thanks, > Zorro > > [1] > SECTION -- default > FSTYP -- xfs (non-debug) > PLATFORM -- Linux/x86_64 dell-per750-41 6.11.0-0.rc6.49.fc42.x86_64+debug #1 SMP PREEMPT_DYNAMIC Mon Sep 2 02:18:15 UTC 2024 > MKFS_OPTIONS -- -f /dev/sda6 > MOUNT_OPTIONS -- -o context=system_u:object_r:root_t:s0 /dev/sda6 /mnt/scratch > > xfs/609 [failed, exit status 1]_check_dmesg: something found in dmesg (see /root/git/xfstests/results//default/xfs/609.dmesg) > - output mismatch (see /root/git/xfstests/results//default/xfs/609.out.bad) > --- tests/xfs/609.out 2024-10-30 16:29:52.250176790 +0800 > +++ /root/git/xfstests/results//default/xfs/609.out.bad 2024-10-30 16:31:01.759590117 +0800 > @@ -1,2 +1,2 @@ > QA output created by 609 > -Silence is golden. > +mkfs.xfs: small data volume, ignoring data volume stripe unit 128 and stripe width 256 > ... > (Run 'diff -u /root/git/xfstests/tests/xfs/609.out /root/git/xfstests/results//default/xfs/609.out.bad' to see the entire diff) > xfs/610 [not run] External volumes not in use, skipped this test > Ran: xfs/609 xfs/610 > Not run: xfs/610 > Failures: xfs/609 > Failed 1 of 2 tests > > > > +_scratch_mount > > + > > +# Grow the filesystem in random sized chunks while stressing and performing > > +# shutdown and recovery. The randomization is intended to create a mix of sub-ag > > +# and multi-ag grows. > > +while [ $size -le $endsize ]; do > > + echo "*** stressing a ${sizeb} block filesystem" >> $seqres.full > > + _stress_scratch > > + incsize=$((RANDOM % 40 * 1048576)) > > + size=`expr $size + $incsize` > > + sizeb=`expr $size / $dbsize` # in data blocks > > + echo "*** growing to a ${sizeb} block filesystem" >> $seqres.full > > + $XFS_GROWFS_PROG -D ${sizeb} $SCRATCH_MNT >> $seqres.full > > + > > + sleep $((RANDOM % 3)) > > + _scratch_shutdown > > + ps -e | grep fsstress > /dev/null 2>&1 > > + while [ $? -eq 0 ]; do > > + $KILLALL_PROG -9 fsstress > /dev/null 2>&1 > > + wait > /dev/null 2>&1 > > + ps -e | grep fsstress > /dev/null 2>&1 > > + done > > + _scratch_cycle_mount || _fail "cycle mount failed" > > +done > /dev/null 2>&1 > > +wait # stop for any remaining stress processes > > + > > +_scratch_unmount > > + > > +echo Silence is golden. > > + > > +status=0 > > +exit > > diff --git a/tests/xfs/609.out b/tests/xfs/609.out > > new file mode 100644 > > index 00000000..8be27d3a > > --- /dev/null > > +++ b/tests/xfs/609.out > > @@ -0,0 +1,2 @@ > > +QA output created by 609 > > +Silence is golden. > > -- > > 2.46.2 > > > > >