On Mon, Nov 18, 2019 at 07:12:12PM +1100, Dave Chinner wrote: > On Fri, Nov 08, 2019 at 01:58:56PM +0800, Shaokun Zhang wrote: > > On 2019/11/7 5:20, Dave Chinner wrote: > > Because percpu_counter_batch was initialized to 256 when there are 128 cpu cores. > > Then we change the agcount=1024, and it also goes to slow path frequently because > > mostly there are no 32768 free inodes. > > Hmmm. I had forgotten the batch size increased with CPU count like > that - I had the thought it was log(ncpus), not linear(ncpus). [.....] > Ok, so maybe we just need a small batch size here, like a value of 4 > or 8 just to avoid every inode alloc/free transaction having to pick > up a global spin lock every time... > > Let me do some testing here... [....] > > Agree, I mean that when delta > 0, there is no need to call percpu_counter_compare in > > xfs_mod_ifree/icount. Ok, so the percpu_counter_sum() only shows up during a create workload here, at ~1.5% of the CPU used. only doing the check when delta < 0 makes no difference to that value. CPU usage of percpu_counter_sum is noise for all other parts of the workload workloads, including the "remove files" part of the benchmark. Hence it is this pattern: > i.e. The pattern we see on inode allocation is: > > icount += 64 > ifree += 64 > ifree-- > .... > ifree-- > icount += 64 > ifree += 64 > ifree-- > .... > ifree-- That is causing the compare CPU usage - all the single decrements are triggering it because we are operating on a new filesystem that has no free inodes other than the cluster we just allocated. Hence avoiding doing the value compare when delta > 0 makes no difference to CPU usage because most of the modifications are subraction. > And on inode freeing, we see the opposite: > > ifree++ > .... > ifree++ > icount -= 64 > ifree -= 64 > ifree++ > .... > ifree++ > icount -= 64 > ifree -= 64 For freeing, we aren't always freeing from the same inode cluster, so the ifree count actually goes up quite a bit before it starts going down as we free clusters. Hence if we keep the batch size small here, we should mostly stay out of the compare path, and the "no compare when delta > 0" will also make a substantial difference here if the ifree count is low. So, I reduced the batch size to 8, and CPU usage during creates dropped from 1.5% to 0.6% on a 16p machine. That's a substantial reduction - if this translates to larger machines, that should bring CPU usage down under 2%.... I was going to send this patch for you to test, but I left this email sitting unsent overnight and I thought about it some more. The reality is, as Christoph said, this ends up being just debug code because we silently ignore the underflow error on production kernels. Hence I think the right thing to do is gut the transaction superblock accounting code so that we need simple asserts and don't bother trying to behave like it's an error we can actually handle sanely. This removes the compare from production kernels completely, so you should see this all go away with the patch below. The difference in performance on my 16p machine is not significant - it is within the normal run-to-run variability of the mdtest benchmark, but the counter compare overhead is gone from the profiles. Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx xfs: gut error handling in xfs_trans_unreserve_and_mod_sb() From: Dave Chinner <dchinner@xxxxxxxxxx> The error handling is only used to fire an assert on debug kernels, so lets get rid of all the complexity and expensive stuff that is done to determine if an assert will fire. Rolling back the changes in the transaction if only one counter underruns them makes all the other counters incorrect, because we still have made that change and are committing the transaction. Hence we can remove all the unwinding, too. And xfs_mod_icount/xfs_mod_ifree are only called from xfs_trans_unreserve_and_mod_sb(), so get rid of them and just directly call the percpu_counter_add/percpu_counter_compare functions. Difference in binary size for a production kernel: Before: text data bss dec hex filename 9486 184 8 9678 25ce fs/xfs/xfs_trans.o 10858 89 12 10959 2acf fs/xfs/xfs_mount.o After: text data bss dec hex filename 8462 184 8 8654 21ce fs/xfs/xfs_trans.o 10510 89 12 10611 2973 fs/xfs/xfs_mount.o So not only does it chop out a lot of source code, it also results in a binary size reduction of ~1.3kB in a very frequently used fast path of the filesystem. Signed-off-by: Dave Chinner <dchinner@xxxxxxxxxx> --- fs/xfs/xfs_mount.c | 33 ------------ fs/xfs/xfs_mount.h | 2 - fs/xfs/xfs_trans.c | 153 +++++++++++++---------------------------------------- 3 files changed, 37 insertions(+), 151 deletions(-) diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c index ba5b6f3b2b88..c59d8f589eb9 100644 --- a/fs/xfs/xfs_mount.c +++ b/fs/xfs/xfs_mount.c @@ -1163,39 +1163,6 @@ xfs_log_sbcount(xfs_mount_t *mp) return xfs_sync_sb(mp, true); } -/* - * Deltas for the inode count are +/-64, hence we use a large batch size - * of 128 so we don't need to take the counter lock on every update. - */ -#define XFS_ICOUNT_BATCH 128 -int -xfs_mod_icount( - struct xfs_mount *mp, - int64_t delta) -{ - percpu_counter_add_batch(&mp->m_icount, delta, XFS_ICOUNT_BATCH); - if (__percpu_counter_compare(&mp->m_icount, 0, XFS_ICOUNT_BATCH) < 0) { - ASSERT(0); - percpu_counter_add(&mp->m_icount, -delta); - return -EINVAL; - } - return 0; -} - -int -xfs_mod_ifree( - struct xfs_mount *mp, - int64_t delta) -{ - percpu_counter_add(&mp->m_ifree, delta); - if (percpu_counter_compare(&mp->m_ifree, 0) < 0) { - ASSERT(0); - percpu_counter_add(&mp->m_ifree, -delta); - return -EINVAL; - } - return 0; -} - /* * Deltas for the block count can vary from 1 to very large, but lock contention * only occurs on frequent small block count updates such as in the delayed diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h index fdb60e09a9c5..0324412238ba 100644 --- a/fs/xfs/xfs_mount.h +++ b/fs/xfs/xfs_mount.h @@ -438,8 +438,6 @@ extern int xfs_initialize_perag(xfs_mount_t *mp, xfs_agnumber_t agcount, xfs_agnumber_t *maxagi); extern void xfs_unmountfs(xfs_mount_t *); -extern int xfs_mod_icount(struct xfs_mount *mp, int64_t delta); -extern int xfs_mod_ifree(struct xfs_mount *mp, int64_t delta); extern int xfs_mod_fdblocks(struct xfs_mount *mp, int64_t delta, bool reserved); extern int xfs_mod_frextents(struct xfs_mount *mp, int64_t delta); diff --git a/fs/xfs/xfs_trans.c b/fs/xfs/xfs_trans.c index f4795fdb7389..7632868bdc92 100644 --- a/fs/xfs/xfs_trans.c +++ b/fs/xfs/xfs_trans.c @@ -527,57 +527,51 @@ xfs_trans_apply_sb_deltas( sizeof(sbp->sb_frextents) - 1); } -STATIC int +static void xfs_sb_mod8( uint8_t *field, int8_t delta) { int8_t counter = *field; + if (!delta) + return; counter += delta; - if (counter < 0) { - ASSERT(0); - return -EINVAL; - } + ASSERT(counter >= 0); *field = counter; - return 0; } -STATIC int +static void xfs_sb_mod32( uint32_t *field, int32_t delta) { int32_t counter = *field; + if (!delta) + return; counter += delta; - if (counter < 0) { - ASSERT(0); - return -EINVAL; - } + ASSERT(counter >= 0); *field = counter; - return 0; } -STATIC int +static void xfs_sb_mod64( uint64_t *field, int64_t delta) { int64_t counter = *field; + if (!delta) + return; counter += delta; - if (counter < 0) { - ASSERT(0); - return -EINVAL; - } + ASSERT(counter >= 0); *field = counter; - return 0; } /* - * xfs_trans_unreserve_and_mod_sb() is called to release unused reservations - * and apply superblock counter changes to the in-core superblock. The + * xfs_trans_unreserve_and_mod_sb() is called to release unused reservations and + * apply superblock counter changes to the in-core superblock. The * t_res_fdblocks_delta and t_res_frextents_delta fields are explicitly NOT * applied to the in-core superblock. The idea is that that has already been * done. @@ -586,7 +580,12 @@ xfs_sb_mod64( * used block counts are not updated in the on disk superblock. In this case, * XFS_TRANS_SB_DIRTY will not be set when the transaction is updated but we * still need to update the incore superblock with the changes. + * + * Deltas for the inode count are +/-64, hence we use a large batch size of 128 + * so we don't need to take the counter lock on every update. */ +#define XFS_ICOUNT_BATCH 128 + void xfs_trans_unreserve_and_mod_sb( struct xfs_trans *tp) @@ -622,20 +621,21 @@ xfs_trans_unreserve_and_mod_sb( /* apply the per-cpu counters */ if (blkdelta) { error = xfs_mod_fdblocks(mp, blkdelta, rsvd); - if (error) - goto out; + ASSERT(!error); } if (idelta) { - error = xfs_mod_icount(mp, idelta); - if (error) - goto out_undo_fdblocks; + percpu_counter_add_batch(&mp->m_icount, idelta, + XFS_ICOUNT_BATCH); + if (idelta < 0) + ASSERT(__percpu_counter_compare(&mp->m_icount, 0, + XFS_ICOUNT_BATCH) >= 0); } if (ifreedelta) { - error = xfs_mod_ifree(mp, ifreedelta); - if (error) - goto out_undo_icount; + percpu_counter_add(&mp->m_ifree, ifreedelta); + if (ifreedelta < 0) + ASSERT(percpu_counter_compare(&mp->m_ifree, 0) >= 0); } if (rtxdelta == 0 && !(tp->t_flags & XFS_TRANS_SB_DIRTY)) @@ -643,95 +643,16 @@ xfs_trans_unreserve_and_mod_sb( /* apply remaining deltas */ spin_lock(&mp->m_sb_lock); - if (rtxdelta) { - error = xfs_sb_mod64(&mp->m_sb.sb_frextents, rtxdelta); - if (error) - goto out_undo_ifree; - } - - if (tp->t_dblocks_delta != 0) { - error = xfs_sb_mod64(&mp->m_sb.sb_dblocks, tp->t_dblocks_delta); - if (error) - goto out_undo_frextents; - } - if (tp->t_agcount_delta != 0) { - error = xfs_sb_mod32(&mp->m_sb.sb_agcount, tp->t_agcount_delta); - if (error) - goto out_undo_dblocks; - } - if (tp->t_imaxpct_delta != 0) { - error = xfs_sb_mod8(&mp->m_sb.sb_imax_pct, tp->t_imaxpct_delta); - if (error) - goto out_undo_agcount; - } - if (tp->t_rextsize_delta != 0) { - error = xfs_sb_mod32(&mp->m_sb.sb_rextsize, - tp->t_rextsize_delta); - if (error) - goto out_undo_imaxpct; - } - if (tp->t_rbmblocks_delta != 0) { - error = xfs_sb_mod32(&mp->m_sb.sb_rbmblocks, - tp->t_rbmblocks_delta); - if (error) - goto out_undo_rextsize; - } - if (tp->t_rblocks_delta != 0) { - error = xfs_sb_mod64(&mp->m_sb.sb_rblocks, tp->t_rblocks_delta); - if (error) - goto out_undo_rbmblocks; - } - if (tp->t_rextents_delta != 0) { - error = xfs_sb_mod64(&mp->m_sb.sb_rextents, - tp->t_rextents_delta); - if (error) - goto out_undo_rblocks; - } - if (tp->t_rextslog_delta != 0) { - error = xfs_sb_mod8(&mp->m_sb.sb_rextslog, - tp->t_rextslog_delta); - if (error) - goto out_undo_rextents; - } - spin_unlock(&mp->m_sb_lock); - return; - -out_undo_rextents: - if (tp->t_rextents_delta) - xfs_sb_mod64(&mp->m_sb.sb_rextents, -tp->t_rextents_delta); -out_undo_rblocks: - if (tp->t_rblocks_delta) - xfs_sb_mod64(&mp->m_sb.sb_rblocks, -tp->t_rblocks_delta); -out_undo_rbmblocks: - if (tp->t_rbmblocks_delta) - xfs_sb_mod32(&mp->m_sb.sb_rbmblocks, -tp->t_rbmblocks_delta); -out_undo_rextsize: - if (tp->t_rextsize_delta) - xfs_sb_mod32(&mp->m_sb.sb_rextsize, -tp->t_rextsize_delta); -out_undo_imaxpct: - if (tp->t_rextsize_delta) - xfs_sb_mod8(&mp->m_sb.sb_imax_pct, -tp->t_imaxpct_delta); -out_undo_agcount: - if (tp->t_agcount_delta) - xfs_sb_mod32(&mp->m_sb.sb_agcount, -tp->t_agcount_delta); -out_undo_dblocks: - if (tp->t_dblocks_delta) - xfs_sb_mod64(&mp->m_sb.sb_dblocks, -tp->t_dblocks_delta); -out_undo_frextents: - if (rtxdelta) - xfs_sb_mod64(&mp->m_sb.sb_frextents, -rtxdelta); -out_undo_ifree: + xfs_sb_mod64(&mp->m_sb.sb_frextents, rtxdelta); + xfs_sb_mod64(&mp->m_sb.sb_dblocks, tp->t_dblocks_delta); + xfs_sb_mod32(&mp->m_sb.sb_agcount, tp->t_agcount_delta); + xfs_sb_mod8(&mp->m_sb.sb_imax_pct, tp->t_imaxpct_delta); + xfs_sb_mod32(&mp->m_sb.sb_rextsize, tp->t_rextsize_delta); + xfs_sb_mod32(&mp->m_sb.sb_rbmblocks, tp->t_rbmblocks_delta); + xfs_sb_mod64(&mp->m_sb.sb_rblocks, tp->t_rblocks_delta); + xfs_sb_mod64(&mp->m_sb.sb_rextents, tp->t_rextents_delta); + xfs_sb_mod8(&mp->m_sb.sb_rextslog, tp->t_rextslog_delta); spin_unlock(&mp->m_sb_lock); - if (ifreedelta) - xfs_mod_ifree(mp, -ifreedelta); -out_undo_icount: - if (idelta) - xfs_mod_icount(mp, -idelta); -out_undo_fdblocks: - if (blkdelta) - xfs_mod_fdblocks(mp, -blkdelta, rsvd); -out: - ASSERT(error == 0); return; }