On 11/20/23 17:35, Yu Kuai wrote:
I'm not sure that change just one queue instead of all queues using the same tag_set won't case any regression, for example, BLK_MQ_F_TAG_QUEUE_SHARED is not cleared, and other queues are still sharing tags fairly while this queue doesn't. Perhaps can we add a helper similiar to __blk_mq_update_nr_hw_queues to update all queues using the same tag_set?
Hi Kuai, How about the patch below? Thanks, Bart. diff --git a/Documentation/ABI/stable/sysfs-block b/Documentation/ABI/stable/sysfs-block index 1fe9a553c37b..7b66eb938882 100644 --- a/Documentation/ABI/stable/sysfs-block +++ b/Documentation/ABI/stable/sysfs-block @@ -269,6 +269,19 @@ Description: specific passthrough mechanisms. +What: /sys/block/<disk>/queue/fair_sharing +Date: November 2023 +Contact: linux-block@xxxxxxxxxxxxxxx +Description: + [RW] If hardware queues are shared across request queues, by + default the request tags are distributed evenly across the + active request queues. If the total number of tags is low and + if the workload differs per request queue this approach may + reduce throughput. This sysfs attribute controls whether or not + the fair tag sharing algorithm is enabled. 1 means enabled + while 0 means disabled. + + What: /sys/block/<disk>/queue/fua Date: May 2018 Contact: linux-block@xxxxxxxxxxxxxxx diff --git a/block/blk-mq-debugfs.c b/block/blk-mq-debugfs.c index 5cbeb9344f2f..f41408103106 100644 --- a/block/blk-mq-debugfs.c +++ b/block/blk-mq-debugfs.c @@ -198,6 +198,7 @@ static const char *const hctx_flag_name[] = { HCTX_FLAG_NAME(NO_SCHED), HCTX_FLAG_NAME(STACKING), HCTX_FLAG_NAME(TAG_HCTX_SHARED), + HCTX_FLAG_NAME(DISABLE_FAIR_TAG_SHARING), }; #undef HCTX_FLAG_NAME diff --git a/block/blk-mq.h b/block/blk-mq.h index f75a9ecfebde..eda6bd0611ea 100644 --- a/block/blk-mq.h +++ b/block/blk-mq.h @@ -416,7 +416,8 @@ static inline bool hctx_may_queue(struct blk_mq_hw_ctx *hctx, { unsigned int depth, users; - if (!hctx || !(hctx->flags & BLK_MQ_F_TAG_QUEUE_SHARED)) + if (!hctx || !(hctx->flags & BLK_MQ_F_TAG_QUEUE_SHARED) || + (hctx->flags & BLK_MQ_F_DISABLE_FAIR_TAG_SHARING)) return true; /* diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c index 0b2d04766324..0009450dc8cf 100644 --- a/block/blk-sysfs.c +++ b/block/blk-sysfs.c @@ -24,6 +24,7 @@ struct queue_sysfs_entry { struct attribute attr; ssize_t (*show)(struct request_queue *, char *); ssize_t (*store)(struct request_queue *, const char *, size_t); + bool no_sysfs_mutex; }; static ssize_t @@ -473,6 +474,55 @@ static ssize_t queue_dax_show(struct request_queue *q, char *page) return queue_var_show(blk_queue_dax(q), page); } +static ssize_t queue_fair_sharing_show(struct request_queue *q, char *page) +{ + struct blk_mq_hw_ctx *hctx; + unsigned long i; + bool fair_sharing = true; + + /* Serialize against blk_mq_realloc_hw_ctxs() */ + mutex_lock(&q->sysfs_lock); + queue_for_each_hw_ctx(q, hctx, i) + if (hctx->flags & BLK_MQ_F_DISABLE_FAIR_TAG_SHARING) + fair_sharing = false; + mutex_unlock(&q->sysfs_lock); + + return sysfs_emit(page, "%u\n", fair_sharing); +} + +static ssize_t queue_fair_sharing_store(struct request_queue *q, + const char *page, size_t count) +{ + const unsigned int DFTS_BIT = ilog2(BLK_MQ_F_DISABLE_FAIR_TAG_SHARING); + struct blk_mq_tag_set *set = q->tag_set; + struct blk_mq_hw_ctx *hctx; + unsigned long i; + int res; + bool val; + + res = kstrtobool(page, &val); + if (res < 0) + return res; + + mutex_lock(&set->tag_list_lock); + clear_bit(DFTS_BIT, &set->flags); + list_for_each_entry(q, &set->tag_list, tag_set_list) { + /* Serialize against blk_mq_realloc_hw_ctxs() */ + mutex_lock(&q->sysfs_lock); + if (val) { + queue_for_each_hw_ctx(q, hctx, i) + clear_bit(DFTS_BIT, &hctx->flags); + } else { + queue_for_each_hw_ctx(q, hctx, i) + set_bit(DFTS_BIT, &hctx->flags); + } + mutex_unlock(&q->sysfs_lock); + } + mutex_unlock(&set->tag_list_lock); + + return count; +} + #define QUEUE_RO_ENTRY(_prefix, _name) \ static struct queue_sysfs_entry _prefix##_entry = { \ .attr = { .name = _name, .mode = 0444 }, \ @@ -486,6 +536,14 @@ static struct queue_sysfs_entry _prefix##_entry = { \ .store = _prefix##_store, \ }; +#define QUEUE_RW_ENTRY_NO_SYSFS_MUTEX(_prefix, _name) \ + static struct queue_sysfs_entry _prefix##_entry = { \ + .attr = { .name = _name, .mode = 0644 }, \ + .show = _prefix##_show, \ + .store = _prefix##_store, \ + .no_sysfs_mutex = true, \ + }; + QUEUE_RW_ENTRY(queue_requests, "nr_requests"); QUEUE_RW_ENTRY(queue_ra, "read_ahead_kb"); QUEUE_RW_ENTRY(queue_max_sectors, "max_sectors_kb"); @@ -542,6 +600,7 @@ QUEUE_RW_ENTRY(queue_nonrot, "rotational"); QUEUE_RW_ENTRY(queue_iostats, "iostats"); QUEUE_RW_ENTRY(queue_random, "add_random"); QUEUE_RW_ENTRY(queue_stable_writes, "stable_writes"); +QUEUE_RW_ENTRY_NO_SYSFS_MUTEX(queue_fair_sharing, "fair_sharing"); #ifdef CONFIG_BLK_WBT static ssize_t queue_var_store64(s64 *var, const char *page) @@ -666,6 +725,7 @@ static struct attribute *blk_mq_queue_attrs[] = { &elv_iosched_entry.attr, &queue_rq_affinity_entry.attr, &queue_io_timeout_entry.attr, + &queue_fair_sharing_entry.attr, #ifdef CONFIG_BLK_WBT &queue_wb_lat_entry.attr, #endif @@ -723,6 +783,10 @@ queue_attr_show(struct kobject *kobj, struct attribute *attr, char *page) if (!entry->show) return -EIO; + + if (entry->no_sysfs_mutex) + return entry->show(q, page); + mutex_lock(&q->sysfs_lock); res = entry->show(q, page); mutex_unlock(&q->sysfs_lock); @@ -741,6 +805,9 @@ queue_attr_store(struct kobject *kobj, struct attribute *attr, if (!entry->store) return -EIO; + if (entry->no_sysfs_mutex) + return entry->store(q, page, length); + mutex_lock(&q->sysfs_lock); res = entry->store(q, page, length); mutex_unlock(&q->sysfs_lock); diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h index 1ab3081c82ed..aadb74aa23a3 100644 --- a/include/linux/blk-mq.h +++ b/include/linux/blk-mq.h @@ -503,7 +503,7 @@ struct blk_mq_tag_set { unsigned int cmd_size; int numa_node; unsigned int timeout; - unsigned int flags; + unsigned long flags; void *driver_data; struct blk_mq_tags **tags; @@ -662,7 +662,8 @@ enum { * or shared hwqs instead of 'mq-deadline'. */ BLK_MQ_F_NO_SCHED_BY_DEFAULT = 1 << 7, - BLK_MQ_F_ALLOC_POLICY_START_BIT = 8, + BLK_MQ_F_DISABLE_FAIR_TAG_SHARING = 1 << 8, + BLK_MQ_F_ALLOC_POLICY_START_BIT = 16, BLK_MQ_F_ALLOC_POLICY_BITS = 1, BLK_MQ_S_STOPPED = 0,