On Thu, Sep 21, 2023 at 11:39:44AM +1000, Dave Chinner wrote: > From: Dave Chinner <dchinner@xxxxxxxxxx> > > fstrim will hold the AGF lock for as long as it takes to walk and > discard all the free space in the AG that meets the userspace trim > criteria. For AGs with lots of free space extents (e.g. millions) > or the underlying device is really slow at processing discard > requests (e.g. Ceph RBD), this means the AGF hold time is often > measured in minutes to hours, not a few milliseconds as we normal > see with non-discard based operations. > > This can result in the entire filesystem hanging whilst the > long-running fstrim is in progress. We can have transactions get > stuck waiting for the AGF lock (data or metadata extent allocation > and freeing), and then more transactions get stuck waiting on the > locks those transactions hold. We can get to the point where fstrim > blocks an extent allocation or free operation long enough that it > ends up pinning the tail of the log and the log then runs out of > space. At this point, every modification in the filesystem gets > blocked. This includes read operations, if atime updates need to be > made. > > To fix this problem, we need to be able to discard free space > extents safely without holding the AGF lock. Fortunately, we already > do this with online discard via busy extents. We can mark free space > extents as "busy being discarded" under the AGF lock and then unlock > the AGF, knowing that nobody will be able to allocate that free > space extent until we remove it from the busy tree. > > Modify xfs_trim_extents to use the same asynchronous discard > mechanism backed by busy extents as is used with online discard. > This results in the AGF only needing to be held for short periods of > time and it is never held while we issue discards. Hence if discard > submission gets throttled because it is slow and/or there are lots > of them, we aren't preventing other operations from being performed > on AGF while we wait for discards to complete... > > Signed-off-by: Dave Chinner <dchinner@xxxxxxxxxx> > --- > fs/xfs/xfs_discard.c | 174 +++++++++++++++++++++++++++++++++++---- > fs/xfs/xfs_extent_busy.c | 34 ++++++-- > fs/xfs/xfs_extent_busy.h | 4 + > 3 files changed, 188 insertions(+), 24 deletions(-) > > diff --git a/fs/xfs/xfs_discard.c b/fs/xfs/xfs_discard.c > index 3f45c7bb94f2..f16b254b5eaa 100644 > --- a/fs/xfs/xfs_discard.c > +++ b/fs/xfs/xfs_discard.c > @@ -19,6 +19,56 @@ > #include "xfs_log.h" > #include "xfs_ag.h" > > +/* > + * Notes on an efficient, low latency fstrim algorithm > + * > + * We need to walk the filesystem free space and issue discards on the free > + * space that meet the search criteria (size and location). We cannot issue > + * discards on extents that might be in use, or are so recently in use they are > + * still marked as busy. To serialise against extent state changes whilst we are > + * gathering extents to trim, we must hold the AGF lock to lock out other > + * allocations and extent free operations that might change extent state. > + * > + * However, we cannot just hold the AGF for the entire AG free space walk whilst > + * we issue discards on each free space that is found. Storage devices can have > + * extremely slow discard implementations (e.g. ceph RBD) and so walking a > + * couple of million free extents and issuing synchronous discards on each > + * extent can take a *long* time. Whilst we are doing this walk, nothing else > + * can access the AGF, and we can stall transactions and hence the log whilst > + * modifications wait for the AGF lock to be released. This can lead hung tasks > + * kicking the hung task timer and rebooting the system. This is bad. > + * > + * Hence we need to take a leaf from the bulkstat playbook. It takes the AGI > + * lock, gathers a range of inode cluster buffers that are allocated, drops the > + * AGI lock and then reads all the inode cluster buffers and processes them. It > + * loops doing this, using a cursor to keep track of where it is up to in the AG > + * for each iteration to restart the INOBT lookup from. > + * > + * We can't do this exactly with free space - once we drop the AGF lock, the > + * state of the free extent is out of our control and we cannot run a discard > + * safely on it in this situation. Unless, of course, we've marked the free > + * extent as busy and undergoing a discard operation whilst we held the AGF > + * locked. > + * > + * This is exactly how online discard works - free extents are marked busy when > + * they are freed, and once the extent free has been committed to the journal, > + * the busy extent record is marked as "undergoing discard" and the discard is > + * then issued on the free extent. Once the discard completes, the busy extent > + * record is removed and the extent is able to be allocated again. > + * > + * In the context of fstrim, if we find a free extent we need to discard, we > + * don't have to discard it immediately. All we need to do it record that free > + * extent as being busy and under discard, and all the allocation routines will > + * now avoid trying to allocate it. Hence if we mark the extent as busy under > + * the AGF lock, we can safely discard it without holding the AGF lock because > + * nothing will attempt to allocate that free space until the discard completes. > + * > + * This also allows us to issue discards asynchronously like we do with online > + * discard, and so for fast devices fstrim will run much faster as we can have > + * multiple discard operations in flight at once, as well as pipeline the free > + * extent search so that it overlaps in flight discard IO. > + */ > + > struct workqueue_struct *xfs_discard_wq; > > static void > @@ -94,21 +144,22 @@ xfs_discard_extents( > } > > > -STATIC int > -xfs_trim_extents( > +static int > +xfs_trim_gather_extents( > struct xfs_perag *pag, > xfs_daddr_t start, > xfs_daddr_t end, > xfs_daddr_t minlen, > + struct xfs_alloc_rec_incore *tcur, > + struct xfs_busy_extents *extents, > uint64_t *blocks_trimmed) > { > struct xfs_mount *mp = pag->pag_mount; > - struct block_device *bdev = mp->m_ddev_targp->bt_bdev; > struct xfs_btree_cur *cur; > struct xfs_buf *agbp; > - struct xfs_agf *agf; > int error; > int i; > + int batch = 100; > > /* > * Force out the log. This means any transactions that might have freed > @@ -120,20 +171,28 @@ xfs_trim_extents( > error = xfs_alloc_read_agf(pag, NULL, 0, &agbp); > if (error) > return error; > - agf = agbp->b_addr; > > cur = xfs_allocbt_init_cursor(mp, NULL, agbp, pag, XFS_BTNUM_CNT); > > /* > - * Look up the longest btree in the AGF and start with it. > + * Look up the extent length requested in the AGF and start with it. > */ > - error = xfs_alloc_lookup_ge(cur, 0, be32_to_cpu(agf->agf_longest), &i); > + if (tcur->ar_startblock == NULLAGBLOCK) > + error = xfs_alloc_lookup_ge(cur, 0, tcur->ar_blockcount, &i); > + else > + error = xfs_alloc_lookup_le(cur, tcur->ar_startblock, > + tcur->ar_blockcount, &i); > if (error) > goto out_del_cursor; > + if (i == 0) { > + /* nothing of that length left in the AG, we are done */ > + tcur->ar_blockcount = 0; > + goto out_del_cursor; > + } > > /* > * Loop until we are done with all extents that are large > - * enough to be worth discarding. > + * enough to be worth discarding or we hit batch limits. > */ > while (i) { > xfs_agblock_t fbno; > @@ -148,7 +207,16 @@ xfs_trim_extents( > error = -EFSCORRUPTED; > break; > } > - ASSERT(flen <= be32_to_cpu(agf->agf_longest)); > + > + if (--batch <= 0) { > + /* > + * Update the cursor to point at this extent so we > + * restart the next batch from this extent. > + */ > + tcur->ar_startblock = fbno; > + tcur->ar_blockcount = flen; > + break; > + } > > /* > * use daddr format for all range/len calculations as that is > @@ -163,6 +231,7 @@ xfs_trim_extents( > */ > if (dlen < minlen) { > trace_xfs_discard_toosmall(mp, pag->pag_agno, fbno, flen); > + tcur->ar_blockcount = 0; > break; > } > > @@ -185,29 +254,98 @@ xfs_trim_extents( > goto next_extent; > } > > - trace_xfs_discard_extent(mp, pag->pag_agno, fbno, flen); > - error = blkdev_issue_discard(bdev, dbno, dlen, GFP_NOFS); > - if (error) > - break; > + xfs_extent_busy_insert_discard(pag, fbno, flen, > + &extents->extent_list); > *blocks_trimmed += flen; > - > next_extent: > error = xfs_btree_decrement(cur, 0, &i); > if (error) > break; > > - if (fatal_signal_pending(current)) { > - error = -ERESTARTSYS; > - break; > - } > + /* > + * If there's no more records in the tree, we are done. Set the > + * cursor block count to 0 to indicate to the caller that there > + * is no more extents to search. > + */ > + if (i == 0) > + tcur->ar_blockcount = 0; > } > > + /* > + * If there was an error, release all the gathered busy extents because > + * we aren't going to issue a discard on them any more. > + */ > + if (error) > + xfs_extent_busy_clear(mp, &extents->extent_list, false); > out_del_cursor: > xfs_btree_del_cursor(cur, error); > xfs_buf_relse(agbp); > return error; > } > > +/* > + * Iterate the free list gathering extents and discarding them. We need a cursor > + * for the repeated iteration of gather/discard loop, so use the longest extent > + * we found in the last batch as the key to start the next. > + */ > +static int > +xfs_trim_extents( > + struct xfs_perag *pag, > + xfs_daddr_t start, > + xfs_daddr_t end, > + xfs_daddr_t minlen, > + uint64_t *blocks_trimmed) > +{ > + struct xfs_alloc_rec_incore tcur = { > + .ar_blockcount = pag->pagf_longest, > + .ar_startblock = NULLAGBLOCK, > + }; > + int error = 0; > + > + do { > + struct xfs_busy_extents *extents; > + > + extents = kzalloc(sizeof(*extents), GFP_KERNEL); > + if (!extents) { > + error = -ENOMEM; > + break; > + } > + > + extents->mount = pag->pag_mount; > + extents->owner = extents; > + INIT_LIST_HEAD(&extents->extent_list); > + > + error = xfs_trim_gather_extents(pag, start, end, minlen, > + &tcur, extents, blocks_trimmed); > + if (error) { > + kfree(extents); > + break; > + } > + > + /* > + * We hand the extent list to the discard function here so the > + * discarded extents can be removed from the busy extent list. > + * This allows the discards to run asynchronously with gathering > + * the next round of extents to discard. > + * > + * However, we must ensure that we do not reference the extent > + * list after this function call, as it may have been freed by > + * the time control returns to us. > + */ > + error = xfs_discard_extents(pag->pag_mount, extents); > + if (error) > + break; > + > + if (fatal_signal_pending(current)) { > + error = -ERESTARTSYS; > + break; > + } > + } while (tcur.ar_blockcount != 0); > + > + return error; > + > +} > + > /* > * trim a range of the filesystem. > * > diff --git a/fs/xfs/xfs_extent_busy.c b/fs/xfs/xfs_extent_busy.c > index 7c2fdc71e42d..746814815b1d 100644 > --- a/fs/xfs/xfs_extent_busy.c > +++ b/fs/xfs/xfs_extent_busy.c > @@ -19,13 +19,13 @@ > #include "xfs_log.h" > #include "xfs_ag.h" > > -void > -xfs_extent_busy_insert( > - struct xfs_trans *tp, > +static void > +xfs_extent_busy_insert_list( > struct xfs_perag *pag, > xfs_agblock_t bno, > xfs_extlen_t len, > - unsigned int flags) > + unsigned int flags, > + struct list_head *busy_list) > { > struct xfs_extent_busy *new; > struct xfs_extent_busy *busyp; > @@ -40,7 +40,7 @@ xfs_extent_busy_insert( > new->flags = flags; > > /* trace before insert to be able to see failed inserts */ > - trace_xfs_extent_busy(tp->t_mountp, pag->pag_agno, bno, len); > + trace_xfs_extent_busy(pag->pag_mount, pag->pag_agno, bno, len); > > spin_lock(&pag->pagb_lock); > rbp = &pag->pagb_tree.rb_node; > @@ -62,10 +62,32 @@ xfs_extent_busy_insert( > rb_link_node(&new->rb_node, parent, rbp); > rb_insert_color(&new->rb_node, &pag->pagb_tree); > > - list_add(&new->list, &tp->t_busy); > + list_add(&new->list, busy_list); > spin_unlock(&pag->pagb_lock); > } > > +void > +xfs_extent_busy_insert( > + struct xfs_trans *tp, > + struct xfs_perag *pag, > + xfs_agblock_t bno, > + xfs_extlen_t len, > + unsigned int flags) > +{ > + xfs_extent_busy_insert_list(pag, bno, len, flags, &tp->t_busy); > +} > + > +void > +xfs_extent_busy_insert_discard( > + struct xfs_perag *pag, > + xfs_agblock_t bno, > + xfs_extlen_t len, > + struct list_head *busy_list) I'm a little surprised that isn't a pointer to a struct xfs_busy_extents object... but I guess I'll ramble about /that/ API in the reply to patch 1. Logic here looks solid enough, Reviewed-by: Darrick J. Wong <djwong@xxxxxxxxxx> --D > +{ > + xfs_extent_busy_insert_list(pag, bno, len, XFS_EXTENT_BUSY_DISCARDED, > + busy_list); > +} > + > /* > * Search for a busy extent within the range of the extent we are about to > * allocate. You need to be holding the busy extent tree lock when calling > diff --git a/fs/xfs/xfs_extent_busy.h b/fs/xfs/xfs_extent_busy.h > index 71c28d031e3b..0639aab336f3 100644 > --- a/fs/xfs/xfs_extent_busy.h > +++ b/fs/xfs/xfs_extent_busy.h > @@ -49,6 +49,10 @@ void > xfs_extent_busy_insert(struct xfs_trans *tp, struct xfs_perag *pag, > xfs_agblock_t bno, xfs_extlen_t len, unsigned int flags); > > +void > +xfs_extent_busy_insert_discard(struct xfs_perag *pag, xfs_agblock_t bno, > + xfs_extlen_t len, struct list_head *busy_list); > + > void > xfs_extent_busy_clear(struct xfs_mount *mp, struct list_head *list, > bool do_discard); > -- > 2.40.1 >