On 1/25/21 8:13 PM, Jeffle Xu wrote: > DM will iterate and poll all polling hardware queues of all target mq > devices when polling IO for dm device. To mitigate the race introduced > by iterating all target hw queues, a per-hw-queue flag is maintained > to indicate whether this polling hw queue currently being polled on or > not. Every polling hw queue is exclusive to one polling instance, i.e., > the polling instance will skip this polling hw queue if this hw queue > currently is being polled by another polling instance, and start > polling on the next hw queue. > > IO polling is enabled when all underlying target devices are capable > of IO polling. The sanity check supports the stacked device model, in > which one dm device may be build upon another dm device. In this case, > the mapped device will check if the underlying dm target device > supports IO polling. > > Signed-off-by: Jeffle Xu <jefflexu@xxxxxxxxxxxxxxxxx> > --- > block/blk-core.c | 8 ++- > drivers/md/dm-core.h | 21 +++++++ > drivers/md/dm-table.c | 127 ++++++++++++++++++++++++++++++++++++++++++ > drivers/md/dm.c | 37 ++++++++++++ > 4 files changed, 192 insertions(+), 1 deletion(-) > > diff --git a/block/blk-core.c b/block/blk-core.c > index 3d93aaa9a49b..d81a5f0faa1d 100644 > --- a/block/blk-core.c > +++ b/block/blk-core.c > @@ -1150,7 +1150,13 @@ int blk_poll(struct request_queue *q, blk_qc_t cookie, bool spin) > struct blk_mq_hw_ctx *hctx = NULL; > struct gendisk *disk = NULL; > > - if (!blk_qc_t_valid(cookie) || > + /* > + * In case of bio-base polling, the returned cookie is actually that of > + * the last split bio. Thus the returned cookie may be BLK_QC_T_NONE, > + * while the previous split bios have already been submitted and queued > + * into the polling hw queue. > + */ > + if ((queue_is_mq(q) && !blk_qc_t_valid(cookie)) || > !test_bit(QUEUE_FLAG_POLL, &q->queue_flags)) > return 0; > > diff --git a/drivers/md/dm-core.h b/drivers/md/dm-core.h > index 086d293c2b03..5a0391aa5cc7 100644 > --- a/drivers/md/dm-core.h > +++ b/drivers/md/dm-core.h > @@ -127,6 +127,24 @@ static inline struct dm_stats *dm_get_stats(struct mapped_device *md) > > #define DM_TABLE_MAX_DEPTH 16 > > +enum target_ctx_type { > + TARGET_TYPE_MQ, > + TARGET_TYPE_DM, > +}; > + > +struct target_ctx { > + union { > + struct { > + struct blk_mq_hw_ctx *hctx; > + struct request_queue *q; > + atomic_t busy; > + }; > + struct mapped_device *md; > + }; > + > + int type; > +}; > + I suddenly realize that this implementation is somehow problematic. This implementation actually buffers hctx of underlying mq device. This can be problematic when the hctx of uderlying mq device can be dynamiclly changed at runtime. For example, nvme RESET command can trigger this issue. Users can send nvme RESET command while there's already one dm device built upon this nvme device. In this case, the hctx map of nvme device will be reallocated when there's one dm device built upon this nvme device. And the original old 'struct blk_mq_hw_ctx *hctx' buffered in the dm layer can be out-of-date. Maybe we could embed 'atomic_t busy' in 'struct blk_mq_hw_ctx', and block layer needs to export one interface iterating all hw queues of given device, just like what I did in the very first RFC version patch: +#define queue_for_each_poll_hw_ctx(q, hctx, i) \ +for ((i) = 0; ((q)->tag_set->nr_maps > HCTX_TYPE_POLL) && \ + (i) < (q)->tag_set->map[HCTX_TYPE_POLL].nr_queues && \ +({ hctx = (q)->queue_hw_ctx[((q)->tag_set->map[HCTX_TYPE_POLL].queue_offset + (i))]; 1; }); \ +(i)++) > struct dm_table { > struct mapped_device *md; > enum dm_queue_mode type; > @@ -162,6 +180,9 @@ struct dm_table { > void *event_context; > > struct dm_md_mempools *mempools; > + > + int num_ctx; > + struct target_ctx *ctxs; > }; > > static inline struct completion *dm_get_completion_from_kobject(struct kobject *kobj) > diff --git a/drivers/md/dm-table.c b/drivers/md/dm-table.c > index 188f41287f18..397bb5f57626 100644 > --- a/drivers/md/dm-table.c > +++ b/drivers/md/dm-table.c > @@ -215,6 +215,8 @@ void dm_table_destroy(struct dm_table *t) > > dm_free_md_mempools(t->mempools); > > + kfree(t->ctxs); > + > kfree(t); > } > > @@ -1194,6 +1196,114 @@ static int dm_table_register_integrity(struct dm_table *t) > return 0; > } > > +static int device_supports_poll(struct dm_target *ti, struct dm_dev *dev, > + sector_t start, sector_t len, void *data) > +{ > + struct request_queue *q = bdev_get_queue(dev->bdev); > + > + return q && test_bit(QUEUE_FLAG_POLL, &q->queue_flags); > +} > + > +static bool dm_table_supports_poll(struct dm_table *t) > +{ > + struct dm_target *ti; > + unsigned int i; > + > + /* Ensure that all targets support iopoll. */ > + for (i = 0; i < dm_table_get_num_targets(t); i++) { > + ti = dm_table_get_target(t, i); > + > + if (!ti->type->iterate_devices || > + !ti->type->iterate_devices(ti, device_supports_poll, NULL)) > + return false; > + } > + > + return true; > +} > + > +static int dm_table_calc_target_ctxs(struct dm_target *ti, > + struct dm_dev *dev, > + sector_t start, sector_t len, > + void *data) > +{ > + int *num = data; > + struct request_queue *q = dev->bdev->bd_disk->queue; > + > + if (queue_is_mq(q)) > + *num += q->tag_set->map[HCTX_TYPE_POLL].nr_queues; > + else > + *num += 1; > + > + return 0; > +} > + > +static int dm_table_fill_target_ctxs(struct dm_target *ti, > + struct dm_dev *dev, > + sector_t start, sector_t len, > + void *data) > +{ > + int *index = data; > + struct target_ctx *ctx; > + struct request_queue *q = dev->bdev->bd_disk->queue; > + > + if (queue_is_mq(q)) { > + int i; > + int num = q->tag_set->map[HCTX_TYPE_POLL].nr_queues; > + int offset = q->tag_set->map[HCTX_TYPE_POLL].queue_offset; > + > + for (i = 0; i < num; i++) { > + ctx = &ti->table->ctxs[(*index)++]; > + ctx->q = q; > + ctx->hctx = q->queue_hw_ctx[offset + i]; > + ctx->type = TARGET_TYPE_MQ; > + /* ctx->busy has been initialized to zero */ > + } > + } else { > + struct mapped_device *md = dev->bdev->bd_disk->private_data; > + > + ctx = &ti->table->ctxs[(*index)++]; > + ctx->md = md; > + ctx->type = TARGET_TYPE_DM; > + } > + > + return 0; > +} > + > +static int dm_table_build_target_ctxs(struct dm_table *t) > +{ > + int i, num = 0, index = 0; > + > + if (!__table_type_bio_based(t->type) || !dm_table_supports_poll(t)) > + return 0; > + > + for (i = 0; i < t->num_targets; i++) { > + struct dm_target *ti = dm_table_get_target(t, i); > + > + if (ti->type->iterate_devices) > + ti->type->iterate_devices(ti, dm_table_calc_target_ctxs, > + &num); > + } > + > + if (WARN_ON(!num)) > + return 0; > + > + t->num_ctx = num; > + > + t->ctxs = kcalloc(num, sizeof(struct target_ctx), GFP_KERNEL); > + if (!t->ctxs) > + return -ENOMEM; > + > + for (i = 0; i < t->num_targets; i++) { > + struct dm_target *ti = dm_table_get_target(t, i); > + > + if (ti->type->iterate_devices) > + ti->type->iterate_devices(ti, dm_table_fill_target_ctxs, > + &index); > + } > + > + return 0; > +} > + > /* > * Prepares the table for use by building the indices, > * setting the type, and allocating mempools. > @@ -1224,6 +1334,10 @@ int dm_table_complete(struct dm_table *t) > if (r) > DMERR("unable to allocate mempools"); > > + r = dm_table_build_target_ctxs(t); > + if (r) > + DMERR("unable to build target hctxs"); > + > return r; > } > > @@ -1883,6 +1997,19 @@ void dm_table_set_restrictions(struct dm_table *t, struct request_queue *q, > #endif > > blk_queue_update_readahead(q); > + > + /* > + * Check for request-based device is remained to > + * dm_mq_init_request_queue()->blk_mq_init_allocated_queue(). > + * Also clear previously set QUEUE_FLAG_POLL* if the new table doesn't > + * support iopoll while reloading table. > + */ > + if (__table_type_bio_based(t->type)) { > + if (t->ctxs) > + q->queue_flags |= QUEUE_FLAG_POLL_MASK; > + else > + q->queue_flags &= ~QUEUE_FLAG_POLL_MASK; > + } > } > > unsigned int dm_table_get_num_targets(struct dm_table *t) > diff --git a/drivers/md/dm.c b/drivers/md/dm.c > index 46ca3b739396..e2be1caa086a 100644 > --- a/drivers/md/dm.c > +++ b/drivers/md/dm.c > @@ -1657,6 +1657,42 @@ static blk_qc_t dm_submit_bio(struct bio *bio) > return BLK_QC_T_NONE; > } > > + > +static int dm_poll_one_md(struct mapped_device *md) > +{ > + int i, num_ctx, srcu_idx, ret = 0; > + struct dm_table *t; > + struct target_ctx *ctxs; > + > + t = dm_get_live_table(md, &srcu_idx); > + num_ctx = t->num_ctx; > + ctxs = t->ctxs; > + > + for (i = 0; i < num_ctx; i++) { > + struct target_ctx *ctx = &ctxs[i]; > + > + if (ctx->type == TARGET_TYPE_MQ) { > + if (!atomic_cmpxchg(&ctx->busy, 0, 1)) { > + ret += blk_mq_poll(ctx->q, ctx->hctx); > + atomic_set(&ctx->busy, 0); > + } > + } else > + ret += dm_poll_one_md(ctx->md); > + } > + > + dm_put_live_table(md, srcu_idx); > + > + return ret; > +} > + > +static int dm_bio_poll(struct request_queue *q, blk_qc_t cookie) > +{ > + struct gendisk *disk = queue_to_disk(q); > + struct mapped_device *md= disk->private_data; > + > + return dm_poll_one_md(md); > +} > + > /*----------------------------------------------------------------- > * An IDR is used to keep track of allocated minor numbers. > *---------------------------------------------------------------*/ > @@ -3049,6 +3085,7 @@ static const struct pr_ops dm_pr_ops = { > }; > > static const struct block_device_operations dm_blk_dops = { > + .iopoll = dm_bio_poll, > .submit_bio = dm_submit_bio, > .open = dm_blk_open, > .release = dm_blk_close, > -- Thanks, Jeffle