On 4/16/21 4:00 PM, Jeffle Xu wrote: > Hi, > How about this patch to remove the extra poll_capable() method? > > And the following 'dm: support IO polling for bio-based dm device' needs > following change. > > ``` > + /* > + * Check for request-based device is remained to > + * dm_mq_init_request_queue()->blk_mq_init_allocated_queue(). > + * For bio-based device, only set QUEUE_FLAG_POLL when all underlying > + * devices supporting polling. > + */ > + if (__table_type_bio_based(t->type)) { > + if (dm_table_supports_poll(t)) { > + blk_queue_flag_set(QUEUE_FLAG_POLL_CAP, q); > + blk_queue_flag_set(QUEUE_FLAG_POLL, q); > + } > + else { > + blk_queue_flag_clear(QUEUE_FLAG_POLL, q); > + blk_queue_flag_clear(QUEUE_FLAG_POLL_CAP, q); > + } > + } > ``` > > > Introduce QUEUE_FLAG_POLL_CAP flag, indicating if the device supports IO > polling or not. Thus both blk-mq and bio-based device could set this > flag at the initialization phase, and then only this flag needs to be > checked instead of rechecking if the device has the ability of IO > polling when enabling IO polling via sysfs. > > For NVMe, the ability of IO polling may change after RESET, since > nvme.poll_queues module parameter may change. Thus the ability of IO > polling need to be rechecked after RESET. > The defect of this approach is that all device drivers that may change tag_set after initialization (e.g., NVMe RESET) need to update QUEUE_FLAG_POLL_CAP. Previous this patch, tag_set is checked directly in queue_poll_store, and thus device drivers don't need to care the queue_flags. > Signed-off-by: Jeffle Xu <jefflexu@xxxxxxxxxxxxxxxxx> > --- > block/blk-mq.c | 5 +++-- > block/blk-sysfs.c | 3 +-- > drivers/nvme/host/core.c | 2 ++ > include/linux/blk-mq.h | 12 ++++++++++++ > include/linux/blkdev.h | 2 ++ > 5 files changed, 20 insertions(+), 4 deletions(-) > > diff --git a/block/blk-mq.c b/block/blk-mq.c > index 414f5d99d9de..55ef6b975169 100644 > --- a/block/blk-mq.c > +++ b/block/blk-mq.c > @@ -3227,9 +3227,10 @@ struct request_queue *blk_mq_init_allocated_queue(struct blk_mq_tag_set *set, > q->tag_set = set; > > q->queue_flags |= QUEUE_FLAG_MQ_DEFAULT; > - if (set->nr_maps > HCTX_TYPE_POLL && > - set->map[HCTX_TYPE_POLL].nr_queues) > + if (blk_mq_poll_capable(set)) { > + blk_queue_flag_set(QUEUE_FLAG_POLL_CAP, q); > blk_queue_flag_set(QUEUE_FLAG_POLL, q); > + } > > q->sg_reserved_size = INT_MAX; > > diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c > index db3268d41274..64f0ab84b606 100644 > --- a/block/blk-sysfs.c > +++ b/block/blk-sysfs.c > @@ -426,8 +426,7 @@ static ssize_t queue_poll_store(struct request_queue *q, const char *page, > unsigned long poll_on; > ssize_t ret; > > - if (!q->tag_set || q->tag_set->nr_maps <= HCTX_TYPE_POLL || > - !q->tag_set->map[HCTX_TYPE_POLL].nr_queues) > + if(!blk_queue_poll_cap(q)) > return -EINVAL; > > ret = queue_var_store(&poll_on, page, count); > diff --git a/drivers/nvme/host/core.c b/drivers/nvme/host/core.c > index bb7da34dd967..5344cc877b05 100644 > --- a/drivers/nvme/host/core.c > +++ b/drivers/nvme/host/core.c > @@ -2210,6 +2210,8 @@ static int nvme_update_ns_info(struct nvme_ns *ns, struct nvme_id_ns *id) > ns->lba_shift = id->lbaf[lbaf].ds; > nvme_set_queue_limits(ns->ctrl, ns->queue); > > + blk_mq_check_poll(ns->disk->queue, ns->disk->queue->tag_set); > + > ret = nvme_configure_metadata(ns, id); > if (ret) > goto out_unfreeze; > diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h > index 2c473c9b8990..ee4c89c8bebc 100644 > --- a/include/linux/blk-mq.h > +++ b/include/linux/blk-mq.h > @@ -618,4 +618,16 @@ blk_qc_t blk_mq_submit_bio(struct bio *bio); > void blk_mq_hctx_set_fq_lock_class(struct blk_mq_hw_ctx *hctx, > struct lock_class_key *key); > > +static inline bool blk_mq_poll_capable(struct blk_mq_tag_set *set) > +{ > + return set->nr_maps > HCTX_TYPE_POLL && > + set->map[HCTX_TYPE_POLL].nr_queues; > +} > + > +static inline void blk_mq_check_poll(struct request_queue *q, > + struct blk_mq_tag_set *set) > +{ > + if (blk_mq_poll_capable(set)) > + blk_queue_flag_set(QUEUE_FLAG_POLL_CAP, q); > +} Sorry it should be > +static inline void blk_mq_check_poll(struct request_queue *q, > + struct blk_mq_tag_set *set) > +{ > + if (blk_mq_poll_capable(set)) > + blk_queue_flag_set(QUEUE_FLAG_POLL_CAP, q); > +} > + else > + blk_queue_flag_clear(QUEUE_FLAG_POLL_CAP, q); > +} > #endif > diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h > index 1e88116dc070..d192a106bf40 100644 > --- a/include/linux/blkdev.h > +++ b/include/linux/blkdev.h > @@ -621,6 +621,7 @@ struct request_queue { > #define QUEUE_FLAG_RQ_ALLOC_TIME 27 /* record rq->alloc_time_ns */ > #define QUEUE_FLAG_HCTX_ACTIVE 28 /* at least one blk-mq hctx is active */ > #define QUEUE_FLAG_NOWAIT 29 /* device supports NOWAIT */ > +#define QUEUE_FLAG_POLL_CAP 30 /* device supports IO polling */ > > #define QUEUE_FLAG_MQ_DEFAULT ((1 << QUEUE_FLAG_IO_STAT) | \ > (1 << QUEUE_FLAG_SAME_COMP) | \ > @@ -668,6 +669,7 @@ bool blk_queue_flag_test_and_set(unsigned int flag, struct request_queue *q); > #define blk_queue_registered(q) test_bit(QUEUE_FLAG_REGISTERED, &(q)->queue_flags) > #define blk_queue_nowait(q) test_bit(QUEUE_FLAG_NOWAIT, &(q)->queue_flags) > #define blk_queue_poll(q) test_bit(QUEUE_FLAG_POLL, &(q)->queue_flags) > +#define blk_queue_poll_cap(q) test_bit(QUEUE_FLAG_POLL_CAP, &(q)->queue_flags) > > extern void blk_set_pm_only(struct request_queue *q); > extern void blk_clear_pm_only(struct request_queue *q); > -- Thanks, Jeffle -- dm-devel mailing list dm-devel@xxxxxxxxxx https://listman.redhat.com/mailman/listinfo/dm-devel