Re: About xfstests generic/361

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, 2019-10-28 at 16:34 -0700, Darrick J. Wong wrote:
> On Mon, Oct 28, 2019 at 05:17:05PM +0800, Ian Kent wrote:
> > Hi Darrick,
> > 
> > Unfortunately I'm having a bit of trouble with my USB keyboard
> > and random key repeats, I lost several important messages this
> > morning due to it.
> > 
> > Your report of the xfstests generic/361 problem was one of them
> > (as was Christoph's mail about the mount code location, I'll post
> > on that a bit later). So I'm going to have to refer to the posts
> > and hope that I can supply enough context to avoid confusion.
> > 
> > Sorry about this.
> > 
> > Anyway, you posted:
> > 
> > "Dunno what's up with this particular patch, but I see regressions
> > on
> > generic/361 (and similar asserts on a few others).  The patches
> > leading
> > up to this patch do not generate this error."
> > 
> > I've reverted back to a point more or less before moving the mount
> > and super block handling code around and tried to reproduce the
> > problem
> > on my test VM and I din't see the problem.
> > 
> > Is there anything I need to do when running the test, other have
> > SCRATCH_MNT and SCRATCH_DEV defined in the local config, and the
> > mount point, and the device existing?
> 
> Um... here's the kernel branch that I used:
> 
> https://git.kernel.org/pub/scm/linux/kernel/git/djwong/xfs-linux.git/log/?h=mount-api-crash

Ok, I'll see what I can do with that.

> 
> Along with:
> 
> MKFS_OPTIONS -- -m crc=0

Right.

> MOUNT_OPTIONS -- -o usrquota,grpquota

It looked like generic/361 used only the SCRATCH_DEV so I thought
that meant making a file system and mounting it within the test.

> 
> and both TEST_DEV and SCRATCH_DEV pointed at boring scsi disks.

My VM disks are VirtIO (file based) virtual disks, so that sounds
a bit different.

Unfortunately I can't use raw disks on the NAS I use for VMs and
I've migrated away from having a desktop machine with a couple of
disks to help with testing.

I have other options if I really need to but it's a little bit
harder to setup and use company lab machines remotely, compared to
local hardware (requesting additional disks is hard to do), and
I'm not sure (probably not) if they can/will use raw disks (or
partitions) either.

> 
> > This could have been a problem with the series I posted because
> > I did have some difficulty resolving some conflicts along the
> > way and may have made mistakes, hence reverting to earlier patches
> > (but also keeping the recent small pre-patch changes).
> 
> Yeah, I had the same problem too; you might spot check the commits in
> there just in case /I/ screwed them up.

I will, yes.

> 
> (I would say 'or rebase on for-next' but (a) I don't know how
> Christoph's mount cleanups intermix with that and (b) let's see if
> this afternoon's for-next is less broken on s390 than this morning's
> was
> <frown>)

I neglected to mention that my series is now based on the for-next
branch as I noticed the get_tree_bdev() fix is present so I can drop
the first patch.

It seemed to me that the for-next branch is the right place to base
the series. I expect there will be the odd bump in the road of course
...

Ian




[Index of Archives]     [XFS Filesystem Development (older mail)]     [Linux Filesystem Development]     [Linux Audio Users]     [Yosemite Trails]     [Linux Kernel]     [Linux RAID]     [Linux SCSI]


  Powered by Linux