On Fri, Mar 05, 2021 at 04:11:32PM +1100, Dave Chinner wrote: > From: Dave Chinner <dchinner@xxxxxxxxxx> > > For every iclog that a CIL push will use up, we need to ensure we > have space reserved for the iclog header in each iclog. It is > extremely difficult to do this accurately with a per-cpu counter > without expensive summing of the counter in every commit. However, > we know what the maximum CIL size is going to be because of the > hard space limit we have, and hence we know exactly how many iclogs > we are going to need to write out the CIL. > > We are constrained by the requirement that small transactions only > have reservation space for a single iclog header built into them. > At commit time we don't know how much of the current transaction > reservation is made up of iclog header reservations as calculated by > xfs_log_calc_unit_res() when the ticket was reserved. As larger > reservations have multiple header spaces reserved, we can steal > more than one iclog header reservation at a time, but we only steal > the exact number needed for the given log vector size delta. > > As a result, we don't know exactly when we are going to steal iclog > header reservations, nor do we know exactly how many we are going to > need for a given CIL. > > To make things simple, start by calculating the worst case number of > iclog headers a full CIL push will require. Record this into an > atomic variable in the CIL. Then add a byte counter to the log > ticket that records exactly how much iclog header space has been > reserved in this ticket by xfs_log_calc_unit_res(). This tells us > exactly how much space we can steal from the ticket at transaction > commit time. > > Now, at transaction commit time, we can check if the CIL has a full > iclog header reservation and, if not, steal the entire reservation > the current ticket holds for iclog headers. This minimises the > number of times we need to do atomic operations in the fast path, > but still guarantees we get all the reservations we need. > > Signed-off-by: Dave Chinner <dchinner@xxxxxxxxxx> > --- > fs/xfs/libxfs/xfs_log_rlimit.c | 2 +- > fs/xfs/libxfs/xfs_shared.h | 3 +- > fs/xfs/xfs_log.c | 12 +++++--- > fs/xfs/xfs_log_cil.c | 55 ++++++++++++++++++++++++++-------- > fs/xfs/xfs_log_priv.h | 20 +++++++------ > 5 files changed, 64 insertions(+), 28 deletions(-) > > diff --git a/fs/xfs/libxfs/xfs_log_rlimit.c b/fs/xfs/libxfs/xfs_log_rlimit.c > index 7f55eb3f3653..75390134346d 100644 > --- a/fs/xfs/libxfs/xfs_log_rlimit.c > +++ b/fs/xfs/libxfs/xfs_log_rlimit.c > @@ -88,7 +88,7 @@ xfs_log_calc_minimum_size( > > xfs_log_get_max_trans_res(mp, &tres); > > - max_logres = xfs_log_calc_unit_res(mp, tres.tr_logres); > + max_logres = xfs_log_calc_unit_res(mp, tres.tr_logres, NULL); This is currently the only call site of xfs_log_calc_unit_res, so if a subsequent patch doesn't make use of that last argument it should go away. (I don't know yet, I haven't looked...) > if (tres.tr_logcount > 1) > max_logres *= tres.tr_logcount; > > diff --git a/fs/xfs/libxfs/xfs_shared.h b/fs/xfs/libxfs/xfs_shared.h > index 8c61a461bf7b..b4791b817fe3 100644 > --- a/fs/xfs/libxfs/xfs_shared.h > +++ b/fs/xfs/libxfs/xfs_shared.h > @@ -48,7 +48,8 @@ extern const struct xfs_buf_ops xfs_symlink_buf_ops; > extern const struct xfs_buf_ops xfs_rtbuf_ops; > > /* log size calculation functions */ > -int xfs_log_calc_unit_res(struct xfs_mount *mp, int unit_bytes); > +int xfs_log_calc_unit_res(struct xfs_mount *mp, int unit_bytes, > + int *niclogs); > int xfs_log_calc_minimum_size(struct xfs_mount *); > > struct xfs_trans_res; > diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c > index 8f4f7ae84358..46a006d41184 100644 > --- a/fs/xfs/xfs_log.c > +++ b/fs/xfs/xfs_log.c > @@ -3312,7 +3312,8 @@ xfs_log_ticket_get( > static int > xlog_calc_unit_res( > struct xlog *log, > - int unit_bytes) > + int unit_bytes, > + int *niclogs) > { > int iclog_space; > uint num_headers; > @@ -3392,15 +3393,18 @@ xlog_calc_unit_res( > /* roundoff padding for transaction data and one for commit record */ > unit_bytes += 2 * log->l_iclog_roundoff; > > + if (niclogs) > + *niclogs = num_headers; > return unit_bytes; > } > > int > xfs_log_calc_unit_res( > struct xfs_mount *mp, > - int unit_bytes) > + int unit_bytes, > + int *niclogs) > { > - return xlog_calc_unit_res(mp->m_log, unit_bytes); > + return xlog_calc_unit_res(mp->m_log, unit_bytes, niclogs); > } > > /* > @@ -3418,7 +3422,7 @@ xlog_ticket_alloc( > > tic = kmem_cache_zalloc(xfs_log_ticket_zone, GFP_NOFS | __GFP_NOFAIL); > > - unit_res = xlog_calc_unit_res(log, unit_bytes); > + unit_res = xlog_calc_unit_res(log, unit_bytes, &tic->t_iclog_hdrs); Ok, so each transaction ticket now gets to know the maximum number of iclog headers that the transaction can consume if we use every last byte of the reservation... > > atomic_set(&tic->t_ref, 1); > tic->t_task = current; > diff --git a/fs/xfs/xfs_log_cil.c b/fs/xfs/xfs_log_cil.c > index 50101336a7f4..f8fb2f59e24c 100644 > --- a/fs/xfs/xfs_log_cil.c > +++ b/fs/xfs/xfs_log_cil.c > @@ -44,9 +44,20 @@ xlog_cil_ticket_alloc( > * transaction overhead reservation from the first transaction commit. > */ > tic->t_curr_res = 0; > + tic->t_iclog_hdrs = 0; > return tic; > } > > +static inline void > +xlog_cil_set_iclog_hdr_count(struct xfs_cil *cil) > +{ > + struct xlog *log = cil->xc_log; > + > + atomic_set(&cil->xc_iclog_hdrs, > + (XLOG_CIL_BLOCKING_SPACE_LIMIT(log) / > + (log->l_iclog_size - log->l_iclog_hsize))); > +} > + > /* > * Unavoidable forward declaration - xlog_cil_push_work() calls > * xlog_cil_ctx_alloc() itself. > @@ -70,6 +81,7 @@ xlog_cil_ctx_switch( > struct xfs_cil *cil, > struct xfs_cil_ctx *ctx) > { > + xlog_cil_set_iclog_hdr_count(cil); ...and I guess every time the CIL gets a fresh context, we also record the maximum number of iclog headers that we might be pushing to disk in one go? Which I guess happens if someone commits a lot of updates to a filesystem, a comitting thread hits the throttle threshold, and now the CIL has to switch contexts and write the old context's transactions to disk? > set_bit(XLOG_CIL_EMPTY, &cil->xc_flags); > ctx->sequence = ++cil->xc_current_sequence; > ctx->cil = cil; > @@ -92,6 +104,7 @@ xlog_cil_init_post_recovery( > { > log->l_cilp->xc_ctx->ticket = xlog_cil_ticket_alloc(log); > log->l_cilp->xc_ctx->sequence = 1; > + xlog_cil_set_iclog_hdr_count(log->l_cilp); > } > > static inline int > @@ -419,7 +432,6 @@ xlog_cil_insert_items( > struct xfs_cil_ctx *ctx = cil->xc_ctx; > struct xfs_log_item *lip; > int len = 0; > - int iclog_space; > int iovhdr_res = 0, split_res = 0, ctx_res = 0; > > ASSERT(tp); > @@ -442,19 +454,36 @@ xlog_cil_insert_items( > test_and_clear_bit(XLOG_CIL_EMPTY, &cil->xc_flags)) > ctx_res = ctx->ticket->t_unit_res; > > - spin_lock(&cil->xc_cil_lock); > - > - /* do we need space for more log record headers? */ > - iclog_space = log->l_iclog_size - log->l_iclog_hsize; > - if (len > 0 && (ctx->space_used / iclog_space != > - (ctx->space_used + len) / iclog_space)) { > - split_res = (len + iclog_space - 1) / iclog_space; > - /* need to take into account split region headers, too */ > - split_res *= log->l_iclog_hsize + sizeof(struct xlog_op_header); > - ctx->ticket->t_unit_res += split_res; > + /* > + * Check if we need to steal iclog headers. atomic_read() is not a > + * locked atomic operation, so we can check the value before we do any > + * real atomic ops in the fast path. If we've already taken the CIL unit > + * reservation from this commit, we've already got one iclog header > + * space reserved so we have to account for that otherwise we risk > + * overrunning the reservation on this ticket. > + * > + * If the CIL is already at the hard limit, we might need more header > + * space that originally reserved. So steal more header space from every > + * commit that occurs once we are over the hard limit to ensure the CIL > + * push won't run out of reservation space. > + * > + * This can steal more than we need, but that's OK. > + */ > + if (atomic_read(&cil->xc_iclog_hdrs) > 0 || If we haven't stolen enough iclog header space... > + ctx->space_used + len >= XLOG_CIL_BLOCKING_SPACE_LIMIT(log)) { ...or we've hit a throttling threshold, in which case we know we're going to push, so we might as well take everything and (I guess?) not give back any reservation that would encourage more commits before we're ready? > + int split_res = log->l_iclog_hsize + > + sizeof(struct xlog_op_header); > + if (ctx_res) > + ctx_res += split_res * (tp->t_ticket->t_iclog_hdrs - 1); > + else > + ctx_res = split_res * tp->t_ticket->t_iclog_hdrs; > + atomic_sub(tp->t_ticket->t_iclog_hdrs, &cil->xc_iclog_hdrs); What happens if xc_iclog_hdrs goes negative? Does that merely mean that we stole more space from the transaction than we needed? Or does it indicate that we're trying to cram too much into a single context? I suppose I worry about what might happen if each transaction's committed items actually somehow eats up every byte of reservation and that actually translates to t_iclog_hdrs iclogs being written out with a particular context, where sum(t_iclog_hdrs) is larger than what xlog_cil_set_iclog_hdr_count() precomputes? --D > } > - tp->t_ticket->t_curr_res -= split_res + ctx_res + len; > - ctx->ticket->t_curr_res += split_res + ctx_res; > + > + spin_lock(&cil->xc_cil_lock); > + tp->t_ticket->t_curr_res -= ctx_res + len; > + ctx->ticket->t_unit_res += ctx_res; > + ctx->ticket->t_curr_res += ctx_res; > ctx->space_used += len; > > /* > diff --git a/fs/xfs/xfs_log_priv.h b/fs/xfs/xfs_log_priv.h > index b0dc3bc9de59..e72d14c76e03 100644 > --- a/fs/xfs/xfs_log_priv.h > +++ b/fs/xfs/xfs_log_priv.h > @@ -140,15 +140,16 @@ enum xlog_iclog_state { > #define XLOG_TIC_LEN_MAX 15 > > typedef struct xlog_ticket { > - struct list_head t_queue; /* reserve/write queue */ > - struct task_struct *t_task; /* task that owns this ticket */ > - xlog_tid_t t_tid; /* transaction identifier : 4 */ > - atomic_t t_ref; /* ticket reference count : 4 */ > - int t_curr_res; /* current reservation in bytes : 4 */ > - int t_unit_res; /* unit reservation in bytes : 4 */ > - char t_ocnt; /* original count : 1 */ > - char t_cnt; /* current count : 1 */ > - char t_flags; /* properties of reservation : 1 */ > + struct list_head t_queue; /* reserve/write queue */ > + struct task_struct *t_task; /* task that owns this ticket */ > + xlog_tid_t t_tid; /* transaction identifier */ > + atomic_t t_ref; /* ticket reference count */ > + int t_curr_res; /* current reservation */ > + int t_unit_res; /* unit reservation */ > + char t_ocnt; /* original count */ > + char t_cnt; /* current count */ > + char t_flags; /* properties of reservation */ > + int t_iclog_hdrs; /* iclog hdrs in t_curr_res */ > } xlog_ticket_t; > > /* > @@ -249,6 +250,7 @@ struct xfs_cil_ctx { > struct xfs_cil { > struct xlog *xc_log; > unsigned long xc_flags; > + atomic_t xc_iclog_hdrs; > struct list_head xc_cil; > spinlock_t xc_cil_lock; > > -- > 2.28.0 >