On 06 Dec 2021 at 11:17, Chandan Babu R wrote: > On 02 Dec 2021 at 21:15, Josef Bacik wrote: >> On Thu, Dec 02, 2021 at 02:45:24PM +0000, Filipe Manana wrote: >>> On Wed, Dec 01, 2021 at 01:18:52PM -0500, Josef Bacik wrote: >>> > We've been seeing transient errors with any test that uses a dm device >>> > for the entirety of the time that we've been running nightly xfstests >>> >>> I have been having it on my tests vms since ever as well. >>> It's really annoying, but fortunatelly it doesn't happen too often. >>> >> >> Yeah before this change we'd fail ~6 tests on every configruation on every >> overnight run. With this change we fail 0. It's rare, but with our continual >> testing it happens sooooooo much. >> >>> > runs. This turns out to be because sometimes we get EBUSY while trying >>> > to create our new dm device. Generally this is because the test comes >>> > right after another test that messes with the dm device, and thus we >>> > still have udev messing around with the device when DM tries to O_EXCL >>> > the block device. >>> > >>> > Add a UDEV_SETTLE_PROG before creating the device to make sure we can >>> > create our new dm device without getting this transient error. >>> >>> I suspect this might only make it seem the problem goes away but does not >>> really fix it. >>> >>> I say that for 2 reasons: >>> >>> 1) All tests that use dm end up calling _dmsetup_remove(), like through >>> _log_writes_remove() or _cleanup_flakey() for example. Normally those >>> are called in the _cleanup() function, which ensures it's done even if >>> the test fails for some reason. >>> >>> So I don't understand why we need that UDEV_SETTLE_PROG at _dmsetup_create(). >>> >>> And I've seen the ebusy failure happen even when the previous tests did >>> not use any dm device; >>> >>> 2) Some tests fail after creating the dm device and using it. For example >>> btrfs/206 often fails when it tries to fsck the filesystem: >>> >>> btrfs/206 3s ... [failed, exit status 1]- output mismatch (see /home/fdmanana/git/hub/xfstests/results//btrfs/206.out.bad) >>> --- tests/btrfs/206.out 2020-10-16 23:13:46.554152652 +0100 >>> +++ /home/fdmanana/git/hub/xfstests/results//btrfs/206.out.bad 2021-12-01 21:09:46.317632589 +0000 >>> @@ -3,3 +3,5 @@ >>> XXX Bytes, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) >>> wrote 8192/8192 bytes at offset 0 >>> XXX Bytes, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec) >>> +_check_btrfs_filesystem: filesystem on /dev/sdc is inconsistent >>> +(see /home/fdmanana/git/hub/xfstests/results//btrfs/206.full for details) >>> ... >>> >>> (Run 'diff -u /home/fdmanana/git/hub/xfstests/tests/btrfs/206.out /home/fdmanana/git/hub/xfstests/results//btrfs/206.out.bad' to see the entire diff) >>> >>> In the .full file I got: >>> >>> (...) >>> replaying 1239@11201: sector 2173408, size 16384, flags 0x10(METADATA) >>> replaying 1240@11234: sector 0, size 0, flags 0x1(FLUSH) >>> replaying 1241@11235: sector 128, size 4096, flags 0x12(FUA|METADATA) >>> _check_btrfs_filesystem: filesystem on /dev/sdc is inconsistent >>> *** fsck.btrfs output *** >>> ERROR: cannot open device '/dev/sdc': Device or resource busy >>> ERROR: cannot open file system >>> Opening filesystem to check... >>> *** end fsck.btrfs output >>> *** mount output *** >>> >>> The ebusy failure is not when the test starts, but when somewhere in the middle >>> of the replay loop when it calls fsck, or when it ends and the fstests framework >>> calls fsck. >>> >>> I've seen that with btrfs/172 too, which also uses dm logwrites in a similar way. >>> >>> So to me this suggests 2 things: >>> >>> 1) Calling UDEV_SETTLE_PROG at _dmsetup_create() doesn't solve that problem with >>> btrfs/206 (and other tests) - the problem is fsck failing to open the scratch >>> device after it called _log_writes_remove() -> _dmsetup_remove(), and not a >>> failure to create the dm device; >>> >>> 2) The problem is likely something missing at _dmsetup_remove(). Perhaps add >>> another UDEV_SETTLE_PROG there: >>> >>> diff --git a/common/rc b/common/rc >>> index 8e351f17..22b34677 100644 >>> --- a/common/rc >>> +++ b/common/rc >>> @@ -4563,6 +4563,7 @@ _dmsetup_remove() >>> $UDEV_SETTLE_PROG >/dev/null 2>&1 >>> $DMSETUP_PROG remove "$@" >>$seqres.full 2>&1 >>> $DMSETUP_PROG mknodes >/dev/null 2>&1 >>> + $UDEV_SETTLE_PROG >/dev/null 2>&1 >>> } >>> >>> _dmsetup_create() >>> >>> I can't say if that change to _dmsetup_remove() is correct, or what it's >>> needed, as I really haven't spent time trying to figure out why the issue >>> happens. >>> >> >> I actually tried a few iterations before I settled on this one, but I was only >> trying to reproduce the EBUSY when creating the flakey device, I hadn't seen it >> with fsck. So I originally started with your change, but it didn't fix the >> problem. Then I did both, UDEV_SETTLE at the end of remove and at the beginning >> of create and the problem went away, and then I removed the one from remove and >> the problem still was gone. >> >> But since I've made this change I also have been investigating another problem >> where we'll get EBUSY at mkfs time when we use SCRATCH_DEV_POOL. I have a test >> patch in our staging branch to make sure it actuall fixes it, but I have to add >> this to the start of _scratch_pool_mkfs as well. >> >> It turns out that udev is doing this thing where it writes to >> /sys/block/whatever/uevent to make sure that a KOBJ_CHANGE event gets sent out >> for the device. >> >> This is racing with the test doing a mount. So the mount gets O_EXCL, which >> means the uevent doesn't get emitted until umount. This would explain what >> you're seeing, we umount, we get the KOBJ_CHANGE event once the O_EXCL is >> dropped, udev runs, and then fsck gets an EBUSY. > > I started debugging this issue late last week. Among several tests which > failed, xfs/033 was failing once in a while because the umount syscall > returned -EBUSY. Debugging this further led me to the following, > > 1. Mounting an fs causes udev to invoke xfs_spaceman (via udisksd => xfs_info). > This causes the per-cpu counter at mount->mnt_pcp->mnt_count to increase by > 1. > 2. Unmount at this stage causes the umount syscall to fail since the refcount > on struct mount is greater than 2. > > I changed my test disks from the regular /dev/sd* devices to loop devices. I > then added loop devices to be ignored by udev with the following change in > /lib/udev/rules.d/60-persistent-storage.rules, > > KERNEL!="sr*|loop*", IMPORT{builtin}="blkid" > > This led to xfs/033 execute fine for 100 times without failure. > However, other tests which use device mapper devices are still failing > arbitrarily. This failure occurs because 60-persistent-storage-dm.rules has the following, # Add inotify watch to track changes on this device. # Using the watch rule is not optimal - it generates a lot of spurious # and useless events whenever the device opened for read-write is closed. # The best would be to generete the event directly in the tool changing # relevant information so only relevant events will be processed # (like creating a filesystem, changing filesystem label etc.). # # But let's use this until we have something better... LABEL="dm_watch" OPTIONS+="watch" Hence any change to the device will generate a uevent causing udev's internal blkid to be invoked. This ends up forking and executing xfs_spaceman. > > Looks like a simple mount operation causes udev to indirectly invoke > xfs_spaceman as described by the output of perf script, > > bash 50034 [002] 2541.601278: sched:sched_process_fork: comm=bash pid=50034 child_comm=bash child_pid=50044 > mount 50044 [003] 2541.601851: sched:sched_process_exec: filename=/usr/bin/mount pid=50044 old_pid=50044 > systemd-udevd 173 [000] 2541.620525: sched:sched_process_fork: comm=systemd-udevd pid=173 child_comm=systemd-udevd child_pid=50048 > systemd-udevd 173 [000] 2541.621071: sched:sched_process_fork: comm=systemd-udevd pid=173 child_comm=systemd-udevd child_pid=50051 > systemd-udevd 173 [000] 2541.621562: sched:sched_process_fork: comm=systemd-udevd pid=173 child_comm=systemd-udevd child_pid=50052 > systemd-udevd 173 [000] 2541.621971: sched:sched_process_fork: comm=systemd-udevd pid=173 child_comm=systemd-udevd child_pid=50053 > systemd-udevd 173 [000] 2541.622657: sched:sched_process_fork: comm=systemd-udevd pid=173 child_comm=systemd-udevd child_pid=50054 > udisksd 588 [000] 2541.675852: sched:sched_process_fork: comm=udisksd pid=588 child_comm=udisksd child_pid=50058 > xfs_admin 50058 [000] 2541.676645: sched:sched_process_exec: filename=/usr/sbin/xfs_admin pid=50058 old_pid=50058 > xfs_admin 50058 [000] 2541.677317: sched:sched_process_fork: comm=xfs_admin pid=50058 child_comm=xfs_admin child_pid=50059 > xfs_db 50059 [000] 2541.677569: sched:sched_process_exec: filename=/usr/sbin/xfs_db pid=50059 old_pid=50059 > udisksd 588 [003] 2541.687075: sched:sched_process_fork: comm=udisksd pid=588 child_comm=udisksd child_pid=50060 > xfs_info 50060 [001] 2541.687843: sched:sched_process_exec: filename=/usr/sbin/xfs_info pid=50060 old_pid=50060 > xfs_info 50060 [001] 2541.688362: sched:sched_process_fork: comm=xfs_info pid=50060 child_comm=xfs_info child_pid=50061 > xfs_info 50061 [003] 2541.688615: sched:sched_process_fork: comm=xfs_info pid=50061 child_comm=xfs_info child_pid=50062 > xfs_info 50062 [001] 2541.688810: sched:sched_process_fork: comm=xfs_info pid=50062 child_comm=xfs_info child_pid=50063 > xfs_info 50062 [001] 2541.688944: sched:sched_process_fork: comm=xfs_info pid=50062 child_comm=xfs_info child_pid=50064 > losetup 50063 [003] 2541.689058: sched:sched_process_exec: filename=/usr/sbin/losetup pid=50063 old_pid=50063 > tail 50064 [000] 2541.689169: sched:sched_process_exec: filename=/usr/bin/tail pid=50064 old_pid=50064 > xfs_info 50060 [001] 2541.690200: sched:sched_process_fork: comm=xfs_info pid=50060 child_comm=xfs_info child_pid=50065 > findmnt 50065 [003] 2541.690501: sched:sched_process_exec: filename=/usr/bin/findmnt pid=50065 old_pid=50065 > xfs_info 50060 [001] 2541.692154: sched:sched_process_fork: comm=xfs_info pid=50060 child_comm=xfs_info child_pid=50066 > xfs_spaceman 50066 [003] 2541.692345: sched:sched_process_exec: filename=/usr/sbin/xfs_spaceman pid=50066 old_pid=50066 > > xfs_spaceman in turn causes ref count on struct mount to increase. Hence an > unmount operation can fail with -EBUSY error. > >> >> This is a very long email to say that udev is causing spurious failures because >> of behavior I don't entirely understand. We're going to need to sprinkle in >> UDEV_SETTLE_PROG in different places to kill all of these different scenarios. >> >> What do we think is best here? Put UDEV_SETTLE_PROG at the start of any >> function that needs to do O_EXCL? So this would mean >> >> _dmsetup_create >> _dmsetup_remove >> *_mkfs >> *_mount >> *_check >> > > To be honest, I don't understand udev well enough. But as pointed above, a > simple invocation of mount is causing xfs_spaceman to be executed (indirectly) > by udevd. Hence, may be executing UDEV_SETTLE_PROG is probably not > sufficient. > >> That would be safest, and I can certainly do that. My initial approach was just >> to do it where it was problematic, but the debugging I did yesterday around >> btrfs/223 failures and the fact that udev is queue'ing up events that get >> delivered at some point in the future makes it kind of hard to handle on a >> case-by-case basis. Thanks, >> >> Josef -- chandan