On Mar 07, 2022 / 20:26, Ming Lei wrote: > On Mon, Mar 07, 2022 at 11:26:56AM +0000, Shinichiro Kawasaki wrote: > > I observed blktests block/005 failure by KASAN use-after-free. The test case > > repeats IO scheduler switch during fio workload. According to the kernel > > message, it looks like bio submit path is running while IO scheduler is invalid, > > and kyber_bio_merge accessed the invalid IO scheduler data. > > > > The failure was observed with v5.17-rc1, and still observed with v5.17-rc7. > > It was observed with QEMU NVME emulation device with ZNS zoned device > > emulation. So far, it is not observed with other devices. To recreate the > > failure, need to repeat the test case. In the worst case, 500 times repetition > > is required. > > > > I bisected and found the commit 9d497e2941c3 ("block: don't protect > > submit_bio_checks by q_usage_counter") triggers it. The commit moves calls of > > 3 functions: submit_bio_checks, blk_mq_attempt_bio_merge and rq_qos_throttle. > > I suspected that the move of blk_mq_attempt_bio_merge is the cause, and tried > > to revert only that part. But it caused linux system boot hang related to > > rq_qos_throttle. Then I created another patch which reverts moves of both > > blk_mq_attempt_bio_merge and rq_qos_throttle calls [2]. With this patch, the > > KASAN uaf disappeared. Based on this observation, I think the failure cause is > > the move of blk_mq_attempt_bio_merge out of guard by bio_queue_enter. > > > > I'm not sure if the fix by the patch [2] is good enough. With that fix, the > > blk_mq_attempt_bio_merge call in blk_mq_get_new_requests is guarded with > > bio_queue_enter, but the blk_mq_attempt_bio_merge call in > > blk_mq_get_cached_request may not be well guarded. Comments for fix approach > > will be appreciated. > > > > > > [1] > > > > [ 335.931534] run blktests block/005 at 2022-03-07 10:15:29 > > [ 336.285062] general protection fault, probably for non-canonical address 0xdffffc0000000011: 0000 [#1] PREEMPT SMP KASAN PTI > > [ 336.291190] KASAN: null-ptr-deref in range [0x0000000000000088-0x000000000000008f] > > [ 336.297513] CPU: 0 PID: 1864 Comm: fio Not tainted 5.16.0-rc3+ #15 > > [ 336.302034] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS rel-1.15.0-29-g6a62e0cb0dfe-prebuilt.qemu.org 04/01/2014 > > [ 336.309636] RIP: 0010:kyber_bio_merge+0x121/0x320 > > I can understand the issue, since both bio merge and rq_qos_throttle() > requires to get one .q_usage_counter, otherwise switching elevator > or changing rq qos may cause such issue. Ming, thanks for the clarification. In the subject, I noted KASAN use-after-free was observed, but the kernel message above shows KASAN null-ptr-deref. Actually, both KASAN use-after-free and null-ptr-deref are observed occasionally, and I mixed them. In case it confused you, sorry about it. > > > > > [2] > > > > diff --git a/block/blk-mq.c b/block/blk-mq.c > > index d69ca91fbc8b..59c66b9e8a44 100644 > > --- a/block/blk-mq.c > > +++ b/block/blk-mq.c > > @@ -2718,7 +2718,8 @@ static bool blk_mq_attempt_bio_merge(struct request_queue *q, > > > > static struct request *blk_mq_get_new_requests(struct request_queue *q, > > struct blk_plug *plug, > > - struct bio *bio) > > + struct bio *bio, > > + unsigned int nsegs) > > { > > struct blk_mq_alloc_data data = { > > .q = q, > > @@ -2730,6 +2731,11 @@ static struct request *blk_mq_get_new_requests(struct request_queue *q, > > if (unlikely(bio_queue_enter(bio))) > > return NULL; > > > > + if (blk_mq_attempt_bio_merge(q, bio, nsegs)) > > + goto queue_exit; > > + > > + rq_qos_throttle(q, bio); > > + > > if (plug) { > > data.nr_tags = plug->nr_ios; > > plug->nr_ios = 1; > > @@ -2742,12 +2748,13 @@ static struct request *blk_mq_get_new_requests(struct request_queue *q, > > rq_qos_cleanup(q, bio); > > if (bio->bi_opf & REQ_NOWAIT) > > bio_wouldblock_error(bio); > > +queue_exit: > > blk_queue_exit(q); > > return NULL; > > } > > > > static inline struct request *blk_mq_get_cached_request(struct request_queue *q, > > - struct blk_plug *plug, struct bio *bio) > > + struct blk_plug *plug, struct bio **bio, unsigned int nsegs) > > { > > struct request *rq; > > > > @@ -2757,14 +2764,20 @@ static inline struct request *blk_mq_get_cached_request(struct request_queue *q, > > if (!rq || rq->q != q) > > return NULL; > > > > - if (blk_mq_get_hctx_type(bio->bi_opf) != rq->mq_hctx->type) > > + if (blk_mq_attempt_bio_merge(q, *bio, nsegs)) { > > + *bio = NULL; > > + return NULL; > > + } > > + > > + if (blk_mq_get_hctx_type((*bio)->bi_opf) != rq->mq_hctx->type) > > return NULL; > > - if (op_is_flush(rq->cmd_flags) != op_is_flush(bio->bi_opf)) > > + if (op_is_flush(rq->cmd_flags) != op_is_flush((*bio)->bi_opf)) > > return NULL; > > > > - rq->cmd_flags = bio->bi_opf; > > + rq->cmd_flags = (*bio)->bi_opf; > > plug->cached_rq = rq_list_next(rq); > > INIT_LIST_HEAD(&rq->queuelist); > > + rq_qos_throttle(q, *bio); > > rq_qos_throttle() can be put above together with blk_mq_attempt_bio_merge(). Okay, will move rq_qos_throttle() just after blk_mq_attempt_bio_merge(). > > > return rq; > > } > > > > @@ -2800,14 +2813,11 @@ void blk_mq_submit_bio(struct bio *bio) > > if (!bio_integrity_prep(bio)) > > return; > > > > - if (blk_mq_attempt_bio_merge(q, bio, nr_segs)) > > - return; > > - > > - rq_qos_throttle(q, bio); > > - > > - rq = blk_mq_get_cached_request(q, plug, bio); > > + rq = blk_mq_get_cached_request(q, plug, &bio, nr_segs); > > if (!rq) { > > - rq = blk_mq_get_new_requests(q, plug, bio); > > + if (!bio) > > + return; > > + rq = blk_mq_get_new_requests(q, plug, bio, nr_segs); > > if (unlikely(!rq)) > > return; > > } > > Otherwise, the patch looks fine, can you post the formal version? Will do. -- Best Regards, Shin'ichiro Kawasaki