On Mon, Mar 19, 2018 at 01:34:47PM -0400, Brian Foster wrote: > On Fri, Mar 16, 2018 at 09:26:02AM -0700, Darrick J. Wong wrote: > > From: Darrick J. Wong <darrick.wong@xxxxxxxxxx> > > > > From the kernel patch that this test examines ("xfs: detect agfl count > > corruption and reset agfl"): > > > > "The struct xfs_agfl v5 header was originally introduced with > > unexpected padding that caused the AGFL to operate with one less > > slot than intended. The header has since been packed, but the fix > > left an incompatibility for users who upgrade from an old kernel > > with the unpacked header to a newer kernel with the packed header > > while the AGFL happens to wrap around the end. The newer kernel > > recognizes one extra slot at the physical end of the AGFL that the > > previous kernel did not. The new kernel will eventually attempt to > > allocate a block from that slot, which contains invalid data, and > > cause a crash. > > > > "This condition can be detected by comparing the active range of the > > AGFL to the count. While this detects a padding mismatch, it can > > also trigger false positives for unrelated flcount corruption. Since > > we cannot distinguish a size mismatch due to padding from unrelated > > corruption, we can't trust the AGFL enough to simply repopulate the > > empty slot. > > > > "Instead, avoid unnecessarily complex detection logic and and use a > > solution that can handle any form of flcount corruption that slips > > through read verifiers: distrust the entire AGFL and reset it to an > > empty state. Any valid blocks within the AGFL are intentionally > > leaked. This requires xfs_repair to rectify (which was already > > necessary based on the state the AGFL was found in). The reset > > mitigates the side effect of the padding mismatch problem from a > > filesystem crash to a free space accounting inconsistency." > > > > This test exercises the reset code by mutating a fresh filesystem to > > contain an agfl with various list configurations of correctly wrapped, > > incorrectly wrapped, not wrapped, and actually corrupt free lists; then > > checks the success of the reset operation by fragmenting the free space > > btrees to exercise the agfl. Kernels without this reset fix will shut > > down the filesystem with corruption errors. > > > > Signed-off-by: Darrick J. Wong <darrick.wong@xxxxxxxxxx> > > --- > > The test fails for me on a kernel with the agfl reset patch: > > # diff -u tests/xfs/709.out /root/xfstests-dev/results//xfs/709.out.bad > --- tests/xfs/709.out 2018-03-19 12:05:20.146356068 -0400 > +++ /root/xfstests-dev/results//xfs/709.out.bad 2018-03-19 > 12:32:01.722735121 -0400 > @@ -7,6 +7,16 @@ > TEST good_start > TEST good_wrap > TEST bad_start > +/mnt/scratch/a: Input/output error > +Error: AG 0 free space header: Repairs are required. > +Error: AG 0 free list: Repairs are required. > +Error: AG 0 freesp by block btree: Repairs are required. > +Error: AG 0 freesp by length btree: Repairs are required. > +Error: AG 0 inode btree: Repairs are required. > +Error: AG 0 free inode btree: Repairs are required. > +Error: dev 253:3 AG 0 fsmap: Structure needs cleaning. > +/mnt/scratch: errors found: 7 > +/mnt/scratch: Unmount and run xfs_repair. > expected flfirst < good_agfl_size - 1 > expected flfirst < fllast > TEST no_move > > I think the errors are expected in this case, but the test probably > shouldn't fail..? (FYI, I didn't have CONFIG_XFS_DEBUG enabled at first. > I tried it again with CONFIG_XFS_DEBUG=y and the test passes.) Hm, that's fun. :) What kernel, xfsprogs, etc.? Or maybe it's easier just to remove the scrub parts until I get that part settled down since in theory the agfl scrubber could just detect the alignment error and return failure immediately. > Some other random comments.. > > > common/rc | 6 + > > tests/xfs/709 | 260 +++++++++++++++++++++++++++++++++++++++++++++++++++++ > > tests/xfs/709.out | 13 +++ > > tests/xfs/group | 1 > > 4 files changed, 280 insertions(+) > > create mode 100755 tests/xfs/709 > > create mode 100644 tests/xfs/709.out > > > > diff --git a/common/rc b/common/rc > > index 2c29d55..8f048f1 100644 > > --- a/common/rc > > +++ b/common/rc > > @@ -3440,6 +3440,12 @@ _get_device_size() > > grep `_short_dev $1` /proc/partitions | awk '{print $3}' > > } > > > > +# check dmesg log for a specific string > > +_check_dmesg_for() { > > + dmesg | tac | sed -ne "0,\#run fstests $seqnum at $date_time#p" | \ > > + tac | egrep -q "$1" > > +} > > + > > # check dmesg log for WARNING/Oops/etc. > > _check_dmesg() > > { > > diff --git a/tests/xfs/709 b/tests/xfs/709 > > new file mode 100755 > > index 0000000..832a469 > > --- /dev/null > > +++ b/tests/xfs/709 > > @@ -0,0 +1,260 @@ > > +#! /bin/bash > > +# FS QA Test No. 709 > > +# > > +# Make sure XFS can fix a v5 AGFL that wraps over the last block. > > +# Refer to commit 96f859d52bcb ("libxfs: pack the agfl header structure so > > +# XFS_AGFL_SIZE is correct") for details on the original on-disk format error > > +# and the patch "xfs: detect agfl count corruption and reset agfl") for details > > +# about the fix. > > +# > > +#----------------------------------------------------------------------- > > +# Copyright (c) 2018 Oracle, Inc. > > +# > > +# This program is free software; you can redistribute it and/or > > +# modify it under the terms of the GNU General Public License as > > +# published by the Free Software Foundation. > > +# > > +# This program is distributed in the hope that it would be useful, > > +# but WITHOUT ANY WARRANTY; without even the implied warranty of > > +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > > +# GNU General Public License for more details. > > +# > > +# You should have received a copy of the GNU General Public License > > +# along with this program; if not, write the Free Software Foundation, > > +# Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA > > +# > > +#----------------------------------------------------------------------- > > +# > > + > > +seq=`basename $0` > > +seqres=$RESULT_DIR/$seq > > +echo "QA output created by $seq" > > + > > +here=`pwd` > > +tmp=/tmp/$$ > > +status=1 > > +trap "_cleanup; rm -f $tmp.*; exit \$status" 0 1 2 3 15 > > + > > +_cleanup() > > +{ > > + cd / > > + rm -f $tmp.* > > +} > > + > > +rm -f $seqres.full > > + > > +# get standard environment, filters and checks > > +. ./common/rc > > +. ./common/xfs > > +. ./common/filter > > +. ./common/populate > > + > > Do we need the xfs and populate includes? common/populate has the helpers for getting/setting metadata object fields. common/xfs has _require_scratch_xfs_crc > > +# real QA test starts here > > +_supported_fs xfs > > +_supported_os Linux > > + > > +_require_freeze > > Or this? Yeah that can go away. > > +_require_scratch_nocheck > > +_require_test_program "punch-alternating" > > + > > +# This is only a v5 filesystem problem > > +_require_scratch_xfs_crc > > + > > +mount_loop() { > > + if ! _try_scratch_mount >> $seqres.full 2>&1; then > > + echo "scratch mount failed" >> $seqres.full > > + return > > + fi > > + > > + # Trigger agfl fixing by fragmenting free space > > + rm -rf $SCRATCH_MNT/a > > + dd if=/dev/zero of=$SCRATCH_MNT/a bs=8192k >> $seqres.full 2>&1 > > I guess we aren't really writing a lot, but fallocate might be more > efficient... > > > + ./src/punch-alternating $SCRATCH_MNT/a > > + sync > > ... and perhaps fsync. Heh, I could skip this entirely since punch-alternating does the fsync for us already. > > + rm -rf $SCRATCH_MNT/a > > + > > + # See if scrub complains... > > + if [ -n "$(_is_mounted $SCRATCH_DEV 2>&1)" ] && \ > > + _supports_xfs_scrub $SCRATCH_MNT $SCRATCH_DEV; then > > + echo "SCRUB" >> $seqres.full > > + "$XFS_SCRUB_PROG" -n $SCRATCH_MNT >> $seqres.full > > + fi > > Is a scrub necessary for the test? Either way, I wonder if this is > something that is better paired with the repair in runtest(). Probably can be omitted for now. > > + > > + _scratch_unmount 2>&1 | _filter_scratch > > +} > > + > > +runtest() { > > + cmd="$1" > > + > > + # Format filesystem > > + echo "TEST $cmd" | tee /dev/ttyprintk > > + echo "TEST $cmd" >> $seqres.full > > + _scratch_unmount >> /dev/null 2>&1 > > + _scratch_mkfs_sized $((32 * 1048576)) >> $seqres.full > > + > > + # Record what was here before > > + echo "FS BEFORE" >> $seqres.full > > + _scratch_xfs_db -c 'sb 0' -c 'p' -c 'agf 0' -c 'p' -c 'agfl 0' -c 'p' > $tmp.before > > + cat $tmp.before >> $seqres.full > > + > > + sectsize=$(_scratch_xfs_get_metadata_field "sectsize" "sb 0") > > + flfirst=$(_scratch_xfs_get_metadata_field "flfirst" "agf 0") > > + fllast=$(_scratch_xfs_get_metadata_field "fllast" "agf 0") > > + flcount=$(_scratch_xfs_get_metadata_field "flcount" "agf 0") > > + > > + # Due to a padding bug in the original v5 struct xfs_agfl, > > + # XFS_AGFL_SIZE could be 36 on 32-bit or 40 on 64-bit. On a system > > + # with 512b sectors, this means that the AGFL length could be > > + # ((512 - 36) / 4) = 119 entries on 32-bit or ((512 - 40) / 4) = 118 > > + # entries on 64-bit. > > + # > > + # We now have code to figure out if the AGFL list wraps incorrectly > > + # according to the kernel's agfl size and fix it by resetting the agfl > > + # to zero length. Mutate ag 0's agfl to be in various configurations > > + # and see if we can trigger the reset. > > + # > > + # Don't hardcode the numbers, calculate them. > > + > > + # Have to have at least three agfl items to test full wrap > > + test "$flcount" -ge 3 || _notrun "insufficient agfl flcount" > > + > > + # mkfs should be able to make us a nice neat flfirst < fllast setup > > + test "$flfirst" -lt "$fllast" || _notrun "fresh agfl already wrapped?" > > + > > + bad_agfl_size=$(( (sectsize - 40) / 4 )) > > + good_agfl_size=$(( (sectsize - 36) / 4 )) > > + agfl_size= > > + case "$1" in > > + "fix_end") # fllast points to the end w/ 40-byte padding > > + new_flfirst=$(( bad_agfl_size - flcount )) > > + agfl_size=$bad_agfl_size;; > > + "fix_start") # flfirst points to the end w/ 40-byte padding > > + new_flfirst=$(( bad_agfl_size - 1)) > > + agfl_size=$bad_agfl_size;; > > + "fix_wrap") # list wraps around end w/ 40-byte padding > > + new_flfirst=$(( bad_agfl_size - (flcount / 2) )) > > + agfl_size=$bad_agfl_size;; > > + "start_zero") # flfirst points to the start > > + new_flfirst=0 > > + agfl_size=$good_agfl_size;; > > + "good_end") # fllast points to the end w/ 36-byte padding > > + new_flfirst=$(( good_agfl_size - flcount )) > > + agfl_size=$good_agfl_size;; > > + "good_start") # flfirst points to the end w/ 36-byte padding > > + new_flfirst=$(( good_agfl_size - 1 )) > > + agfl_size=$good_agfl_size;; > > + "good_wrap") # list wraps around end w/ 36-byte padding > > + new_flfirst=$(( good_agfl_size - (flcount / 2) )) > > + agfl_size=$good_agfl_size;; > > + "bad_start") # flfirst points off the end > > + new_flfirst=$good_agfl_size > > + agfl_size=$good_agfl_size;; > > + "no_move") # whatever mkfs formats (flfirst points to start) > > + new_flfirst=$flfirst > > + agfl_size=$good_agfl_size;; > > + "simple_move") # move list arbitrarily > > + new_flfirst=$((fllast + 1)) > > + agfl_size=$good_agfl_size;; > > + *) > > + _fail "Internal test error";; > > + esac > > + new_fllast=$(( (new_flfirst + flcount - 1) % agfl_size )) > > + > > + # Log what we're doing... > > + cat >> $seqres.full << ENDL > > +sector size: $sectsize > > +bad_agfl_size: $bad_agfl_size [0 - $((bad_agfl_size - 1))] > > +good_agfl_size: $good_agfl_size [0 - $((good_agfl_size - 1))] > > +agfl_size: $agfl_size > > +flfirst: $flfirst > > +fllast: $fllast > > +flcount: $flcount > > +new_flfirst: $new_flfirst > > +new_fllast: $new_fllast > > +ENDL > > + > > + # Remap the agfl blocks > > + echo "$((good_agfl_size - 1)) 0xffffffff" > $tmp.remap > > + seq "$flfirst" "$fllast" | while read f; do > > + list_pos=$((f - flfirst)) > > + dest_pos=$(( (new_flfirst + list_pos) % agfl_size )) > > + bno=$(_scratch_xfs_get_metadata_field "bno[$f]" "agfl 0") > > + echo "$dest_pos $bno" >> $tmp.remap > > + done > > + > > + cat $tmp.remap | while read dest_pos bno junk; do > > + _scratch_xfs_set_metadata_field "bno[$dest_pos]" "$bno" "agfl 0" >> $seqres.full > > + done > > + > > Might be worth factoring the above into a function. Also, do we need all > of the $seqres.full redirection if we dump the $tmp.corrupt bits right > after? Probably not, but I like to preserve the log of what xfs_db did vs. what ended up on disk just to confirm that the _scratch_xfs_set_metadata_field are behaving like they're supposed to. > > + # Set new flfirst/fllast > > + _scratch_xfs_set_metadata_field "fllast" "$new_fllast" "agf 0" >> $seqres.full > > + _scratch_xfs_set_metadata_field "flfirst" "$new_flfirst" "agf 0" >> $seqres.full > > + > > + echo "FS AFTER" >> $seqres.full > > + _scratch_xfs_db -c 'sb 0' -c 'p' -c 'agf 0' -c 'p' -c 'agfl 0' -c 'p' > $tmp.corrupt 2> /dev/null > > + diff -u $tmp.before $tmp.corrupt >> $seqres.full > > + > > + # Mount and see what happens > > + mount_loop > > + > > + # Did we end up with a non-wrapped list? > > + flfirst=$(_scratch_xfs_get_metadata_field "flfirst" "agf 0" 2>/dev/null) > > + fllast=$(_scratch_xfs_get_metadata_field "fllast" "agf 0" 2>/dev/null) > > + if [ "${flfirst}" -ge "$((good_agfl_size - 1))" ]; then > > + echo "expected flfirst < good_agfl_size - 1" > > + echo "expected flfirst(${flfirst}) < good_agfl_size - 1($((good_agfl_size - 1)))" >> $seqres.full > > + fi > > + if [ "${fllast}" -ge "$((good_agfl_size - 1))" ]; then > > + echo "expected fllast < good_agfl_size - 1" > > + echo "expected fllast(${fllast}) < good_agfl_size - 1($((good_agfl_size - 1)))" >> $seqres.full > > + fi > > + if [ "${flfirst}" -ge "${fllast}" ]; then > > + echo "expected flfirst < fllast" > > + echo "expected flfirst(${flfirst}) < fllast(${fllast})" >> $seqres.full > > + fi > > Might be able to use tee here to avoid some of the echo duplication. It > looks like we already dump the raw agf/agfl structures to $seqres.full > below. > > Also note that there are a bunch of lines beyond 80 chars. > > > + > > + echo "FS MOUNTLOOP" >> $seqres.full > > + _scratch_xfs_db -c 'sb 0' -c 'p' -c 'agf 0' -c 'p' -c 'agfl 0' -c 'p' > $tmp.mountloop 2> /dev/null > > + diff -u $tmp.corrupt $tmp.mountloop >> $seqres.full > > + > > + # Let's see what repair thinks > > + echo "REPAIR" >> $seqres.full > > + _scratch_xfs_repair >> $seqres.full 2>&1 > > I guess we don't need _require_scratch_nocheck if we repair before the > test returns. Yep. > > + > > + echo "FS REPAIR" >> $seqres.full > > + _scratch_xfs_db -c 'sb 0' -c 'p' -c 'agf 0' -c 'p' -c 'agfl 0' -c 'p' > $tmp.repair 2> /dev/null > > + diff -u $tmp.mountloop $tmp.repair >> $seqres.full > > + > > + # Try mount/unmount one more time. > > + mount_loop > > + > > + echo "FS REMOUNT" >> $seqres.full > > + _scratch_xfs_db -c 'sb 0' -c 'p' -c 'agf 0' -c 'p' -c 'agfl 0' -c 'p' > $tmp.remount 2> /dev/null > > + diff -u $tmp.repair $tmp.remount >> $seqres.full > > These last couple of hunks seem superfluous. What's the purpose, just to > work out the fs some more? I suppose that makes sense. The comment could > be made more clear. # Exercise the filesystem again to make sure there aren't any lasting # ill effects from either the agfl reset or the recommended subsequent # repair run. > > +} > > + > > +runtest fix_end > > +runtest fix_start > > +runtest fix_wrap > > +runtest start_zero > > +runtest good_end > > +runtest good_start > > +runtest good_wrap > > +runtest bad_start > > +runtest no_move > > +runtest simple_move > > + > > +_scratch_unmount >> $seqres.full 2>&1 > > + > > The scratch mounting/unmounting seems unbalanced. runtest() unmounts the > fs at the start, but it doesn't appear we ever call it with scratch > already mounted. The mount loop cycles the mount, so it seems it should > already be unmounted by the time we get here as well. I think that's a desperate last gasp attempt to scrape the fs off the system when I was working on my version of the patch. It can go away. Thanks for the review! :) --D > Brian > > > +# Did we get the kernel warning too? > > +warn_str='WARNING: Reset corrupted AGFL' > > +_check_dmesg_for "${warn_str}" || echo "Missing kernel log message \"${warn_str}\"." > > + > > +# Now run the regular dmesg check, filtering out the agfl warning > > +filter_agfl_reset_printk() { > > + grep -v "${warn_str}" > > +} > > +_check_dmesg filter_agfl_reset_printk > > + > > +status=0 > > +exit 0 > > diff --git a/tests/xfs/709.out b/tests/xfs/709.out > > new file mode 100644 > > index 0000000..980b4d1 > > --- /dev/null > > +++ b/tests/xfs/709.out > > @@ -0,0 +1,13 @@ > > +QA output created by 709 > > +TEST fix_end > > +TEST fix_start > > +TEST fix_wrap > > +TEST start_zero > > +TEST good_end > > +TEST good_start > > +TEST good_wrap > > +TEST bad_start > > +expected flfirst < good_agfl_size - 1 > > +expected flfirst < fllast > > +TEST no_move > > +TEST simple_move > > diff --git a/tests/xfs/group b/tests/xfs/group > > index e2397fe..472120e 100644 > > --- a/tests/xfs/group > > +++ b/tests/xfs/group > > @@ -441,3 +441,4 @@ > > 441 auto quick clone quota > > 442 auto stress clone quota > > 443 auto quick ioctl fsr > > +709 auto quick > > -- > > To unsubscribe from this list: send the line "unsubscribe fstests" in > > the body of a message to majordomo@xxxxxxxxxxxxxxx > > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- > To unsubscribe from this list: send the line "unsubscribe fstests" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html