On Thu, Nov 01, 2018 at 12:32:59PM +0530, Chandan Rajendra wrote: > On Wednesday, October 31, 2018 5:41:11 PM IST Brian Foster wrote: > > On Tue, Oct 23, 2018 at 12:18:08PM +0530, Chandan Rajendra wrote: > > > generic/305 fails on a 64k block sized filesystem due to the following > > > interaction, > > > > > > 1. We are writing 8 blocks (i.e. [0, 512k-1]) of data to a 1 MiB file. > > > 2. XFS reserves 32 blocks of space in the CoW fork. > > > xfs_bmap_extsize_align() calculates XFS_DEFAULT_COWEXTSZ_HINT (32 > > > blocks) as the number of blocks to be reserved. > > > 3. The reserved space in the range [1M(i.e. i_size), 1M + 16 > > > blocks] is freed by __fput(). This corresponds to freeing "eof > > > blocks" i.e. space reserved beyond EOF of a file. > > > > > > > This still refers to the COW fork, right? > > Yes, xfs_itruncate_extents_flags() invokes xfs_reflink_cancel_cow_blocks() > when "data fork" is being truncated. > > > > > > The reserved space to which data was never written i.e. [9th block, > > > 1M(EOF)], remains reserved in the CoW fork until either the CoW block > > > reservation trimming worker gets invoked or the filesystem is > > > unmounted. > > > > > > > And so this refers to cowblocks within EOF..? If so, that means those > > blocks are consumed if that particular range of the file is written as > > well. The above sort of reads like they'd stick around without any real > > purpose, which is either a bit confusing or suggests I'm missing > > something. > > Yes, the above mentioned range (within inode->i_isize) does not have any data > written to. The space was speculatively reserved. > Sure, that might be true of the test case, but the purpose of allocation hint is essentially to speculate on future writes. Without it, a set of small and scattered writes over a range of shared blocks in a file results in about equally as many small allocations and can fragment the file. > > > > This also all sounds like expected behavior to this point.. > > > > > This commit fixes the issue by freeing unused CoW block reservations > > > whenever quota numbers are requested by userspace application. > > > > > > > Could you elaborate more on the fundamental problem wrt to quota? Are > > the cow blocks not accounted properly or something? What exactly makes > > this a problem with 64k page sizes and not the more common 4k page/block > > size? > > The speculative allocation of CoW blocks are in units of blocks. The default > CoW extent size hint is set to XFS_DEFAULT_COWEXTSZ_HINT (i.e. 32 blocks). For > 4k block size this equals 131072 bytes while for 64k block size it is 2097152 > bytes. > > generic/305 initially creates 1MiB file. It then creates another file which > shares its data blocks with the original file. The test then writes 512K worth > of data at file range [0, 512k-1]. Now here is where we have a difference b/w > 4k v/s 64k block sized filesystems. > Ok.. > Writing 512k data causes max(data written, 32 blocks) of space to be reserved > in the CoW fork i.e 512k bytes for 4k block FS and 2097152 bytes for 64k block > FS. On 4k block FS, the reservation in CoW fork gets cleared when 512k bytes > of data are written to disk. However for 64k block FS, 2097152 - 512k = > 1572864 bytes remain in CoW fork until either the CoW space trimming worker > gets triggered or until the filesystem is umounted. > Yep, but this strikes me as an implementation detail of the test. IOW, if the test issued a smaller write that didn't fully consume the 32-block allocation hint with 4k blocks, we'd be in the same state. So this patch implies that there's some kind of problem with quota stats/reporting with active COW fork reservations but doesn't actually explain what it is. > > > > > > Signed-off-by: Chandan Rajendra <chandan@xxxxxxxxxxxxxxxxxx> > > > --- > > > > > > PS: With the above patch, the tests xfs/214 & xfs/440 fail because the > > > value passed to xfs_io's cowextsize does not have any effect when CoW > > > fork reservations are flushed before querying for quota usage numbers. > > > > > > fs/xfs/xfs_quotaops.c | 13 +++++++++++++ > > > 1 file changed, 13 insertions(+) > > > > > > diff --git a/fs/xfs/xfs_quotaops.c b/fs/xfs/xfs_quotaops.c > > > index a7c0c65..9236a38 100644 > > > --- a/fs/xfs/xfs_quotaops.c > > > +++ b/fs/xfs/xfs_quotaops.c > > > @@ -218,14 +218,21 @@ xfs_fs_get_dqblk( > > > struct kqid qid, > > > struct qc_dqblk *qdq) > > > { > > > + int ret; > > > struct xfs_mount *mp = XFS_M(sb); > > > xfs_dqid_t id; > > > + struct xfs_eofblocks eofb = { 0 }; > > > > > > if (!XFS_IS_QUOTA_RUNNING(mp)) > > > return -ENOSYS; > > > if (!XFS_IS_QUOTA_ON(mp)) > > > return -ESRCH; > > > > > > + eofb.eof_flags = XFS_EOF_FLAGS_SYNC; > > > + ret = xfs_icache_free_cowblocks(mp, &eofb); > > > + if (ret) > > > + return ret; > > > + > > > > So this is a full scan of the in-core icache per call. I'm not terribly > > familiar with the quota infrastructure code, but just from the context > > it looks like this is per quota id. The eofblocks infrastructure > > supports id filtering, which makes me wonder (at minimum) why we > > wouldn't limit the scan to the id associated with the quota? > > I now think replacing the call to "$XFS_SPACEMAN_PROG -c 'prealloc -s' call" > in _check_quota_usage() with umount/mount cycle is the right thing to do. > Ok. Sounds like it's a test issue one way or another then... Brian > Quoting my response to Darrick's mail, > > ;; Hmm. W.r.t Preallocated EOF blocks, it is easy to identify the blocks to be > ;; removed by the ioctl i.e. blocks which are present beyond inode->i_size. > > ;; You are right about the inability to do so for CoW blocks since some of the > ;; unused CoW blocks fall within inode->i_size. Hence I agree with your approach > ;; of replacing "$XFS_SPACEMAN_PROG -c 'prealloc -s' call' in _check_quota_usage > ;; with umount/mount. > > > > > Brian > > > > > id = from_kqid(&init_user_ns, qid); > > > return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq); > > > } > > > @@ -240,12 +247,18 @@ xfs_fs_get_nextdqblk( > > > int ret; > > > struct xfs_mount *mp = XFS_M(sb); > > > xfs_dqid_t id; > > > + struct xfs_eofblocks eofb = { 0 }; > > > > > > if (!XFS_IS_QUOTA_RUNNING(mp)) > > > return -ENOSYS; > > > if (!XFS_IS_QUOTA_ON(mp)) > > > return -ESRCH; > > > > > > + eofb.eof_flags = XFS_EOF_FLAGS_SYNC; > > > + ret = xfs_icache_free_cowblocks(mp, &eofb); > > > + if (ret) > > > + return ret; > > > + > > > id = from_kqid(&init_user_ns, *qid); > > > ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type), > > > qdq); > > > > > > > -- > chandan >