blk_mq_hctx_next_cpu() implements a form of cpumask_next_and_wrap() using cpumask_next_and_cpu() and blk_mq_first_mapped_cpu(): [ 5.398453] WARNING: CPU: 3 PID: 162 at include/linux/cpumask.h:110 __blk_mq_delay_run_hw_queue+0x16b/0x180 [ 5.399317] Modules linked in: [ 5.399646] CPU: 3 PID: 162 Comm: ssh-keygen Tainted: G N 6.0.0-rc4-00004-g93003cb24006 #55 [ 5.400135] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.14.0-0-g155821a1990b-prebuilt.qemu.org 04/01/2014 [ 5.405430] Call Trace: [ 5.406152] <TASK> [ 5.406452] blk_mq_sched_insert_requests+0x67/0x150 [ 5.406759] blk_mq_flush_plug_list+0xd0/0x280 [ 5.406987] ? bit_wait+0x60/0x60 [ 5.407317] __blk_flush_plug+0xdb/0x120 [ 5.407561] ? bit_wait+0x60/0x60 [ 5.407765] io_schedule_prepare+0x38/0x40 [...] This triggers a warning when next_cpu == nr_cpu_ids - 1, so rewrite it using cpumask_next_and_wrap() directly. The backwards-going goto can be removed, as the cpumask_next*() operation already ANDs hctx->cpumask and cpu_online_mask, which implies checking for an online CPU. No change in behaviour intended. Suggested-by: Yury Norov <yury.norov@xxxxxxxxx> Signed-off-by: Valentin Schneider <vschneid@xxxxxxxxxx> --- block/blk-mq.c | 39 +++++++++++++-------------------------- 1 file changed, 13 insertions(+), 26 deletions(-) diff --git a/block/blk-mq.c b/block/blk-mq.c index c96c8c4f751b..1520794dd9ea 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -2038,42 +2038,29 @@ static inline int blk_mq_first_mapped_cpu(struct blk_mq_hw_ctx *hctx) */ static int blk_mq_hctx_next_cpu(struct blk_mq_hw_ctx *hctx) { - bool tried = false; int next_cpu = hctx->next_cpu; if (hctx->queue->nr_hw_queues == 1) return WORK_CPU_UNBOUND; - if (--hctx->next_cpu_batch <= 0) { -select_cpu: - next_cpu = cpumask_next_and(next_cpu, hctx->cpumask, - cpu_online_mask); - if (next_cpu >= nr_cpu_ids) - next_cpu = blk_mq_first_mapped_cpu(hctx); + if (--hctx->next_cpu_batch > 0 && cpu_online(next_cpu)) + return next_cpu; + + next_cpu = cpumask_next_and_wrap(next_cpu, hctx->cpumask, cpu_online_mask, next_cpu, false); + if (next_cpu < nr_cpu_ids) { hctx->next_cpu_batch = BLK_MQ_CPU_WORK_BATCH; + hctx->next_cpu = next_cpu; + return next_cpu; } /* - * Do unbound schedule if we can't find a online CPU for this hctx, - * and it should only happen in the path of handling CPU DEAD. + * No other online CPU in hctx->cpumask. + * + * Make sure to re-select CPU next time once after CPUs + * in hctx->cpumask become online again. */ - if (!cpu_online(next_cpu)) { - if (!tried) { - tried = true; - goto select_cpu; - } - - /* - * Make sure to re-select CPU next time once after CPUs - * in hctx->cpumask become online again. - */ - hctx->next_cpu = next_cpu; - hctx->next_cpu_batch = 1; - return WORK_CPU_UNBOUND; - } - - hctx->next_cpu = next_cpu; - return next_cpu; + hctx->next_cpu_batch = 1; + return WORK_CPU_UNBOUND; } /** -- 2.31.1