On Thu, Sep 19, 2019 at 09:09:33AM +0000, Damien Le Moal wrote: > On 2019/09/19 10:56, Ming Lei wrote: > > On Thu, Sep 19, 2019 at 08:26:32AM +0000, Damien Le Moal wrote: > >> On 2019/09/18 18:30, Alan Stern wrote: > >>> On Wed, 18 Sep 2019, Andrea Vai wrote: > >>> > >>>>> Also, I wonder if the changing the size of the data transfers would > >>>>> make any difference. This is easy to try; just write "64" to > >>>>> /sys/block/sd?/queue/max_sectors_kb (where the ? is the appropriate > >>>>> drive letter) after the drive is plugged in but before the test > >>>>> starts. > >>>> > >>>> ok, so I duplicated the tests above for the "64" case (it was > >>>> initially set as "120", if it is relevant to know), leading to 40 tests named as > >>>> > >>>> bad.mon.out_50000000_64_TIMESTAMP > >>>> bad.mon.out_50000000_non64_TIMESTAMP > >>>> good.mon.out_50000000_64_TIMESTAMP > >>>> good.mon.out_50000000_non64_TIMESTAMP > >>>> > >>>> where "64" denotes the ones done with that value in max_sectors_kb, > >>>> and "not64" the ones without it (as far as I can tell, it has been > >>>> always "120"). > >>>> > >>>> So, we have 40 traces total. Each set of 10 trials is identified by > >>>> a text file, which contains the output log of the test script (and the > >>>> timestamps), also available in the download zipfile. > >>>> > >>>> Just to summarize here the times, they are respectively (number > >>>> expressed in seconds): > >>>> > >>>> BAD: > >>>> Logs: log_10trials_50MB_BAD_NonCanc_64.txt, > >>>> log_10trials_50MB_BAD_NonCanc_non64.txt > >>>> 64: 34, 34, 35, 39, 37, 32, 42, 44, 43, 40 > >>>> not64: 61, 71, 59, 71, 62, 75, 62, 70, 62, 68 > >>>> GOOD: > >>>> Logs: log_10trials_50MB_GOOD_NonCanc_64.txt, > >>>> log_10trials_50MB_GOOD_NonCanc_non64.txt > >>>> 64: 34, 32, 35, 34, 35, 33, 34, 33, 33, 33 > >>>> not64: 32, 30, 32, 31, 31, 30, 32, 30, 32, 31 > >>> > >>> The improvement from using "64" with the bad kernel is quite large. > >>> That alone would be a big help for you. > >>> > >>> However, I did see what appears to be a very significant difference > >>> between the bad and good kernel traces. It has to do with the order in > >>> which the blocks are accessed. > >>> > >>> Here is an extract from one of the bad traces. I have erased all the > >>> information except for the columns containing the block numbers to be > >>> written: > >>> > >>> 00019628 00 > >>> 00019667 00 > >>> 00019628 80 > >>> 00019667 80 > >>> 00019629 00 > >>> 00019668 00 > >>> 00019629 80 > >>> 00019668 80 > >>> > >>> Here is the equivalent portion from one of the good traces: > >>> > >>> 00019628 00 > >>> 00019628 80 > >>> 00019629 00 > >>> 00019629 80 > >>> 0001962a 00 > >>> 0001962a 80 > >>> 0001962b 00 > >>> 0001962b 80 > >>> > >>> Notice that under the good kernel, the block numbers increase > >>> monotonically in a single sequence. But under the bad kernel, the > >>> block numbers are not monotonic -- it looks like there are two separate > >>> threads each with its own strictly increasing sequence. > >>> > >>> This is exactly the sort of difference one might expect to see from > >>> the commit f664a3cc17b7 ("scsi: kill off the legacy IO path") you > >>> identified as the cause of the problem. With multiqueue I/O, it's not > >>> surprising to see multiple sequences of block numbers. > >>> > >>> Add it's not at all surprising that a consumer-grade USB storage device > >>> might do a much worse job of handling non-sequential writes than > >>> sequential ones. > >>> > >>> Which leads to a simple question for the SCSI or block-layer > >>> maintainers: Is there a sysfs setting Andrea can tweak which will > >>> effectively restrict a particular disk device down to a single I/O > >>> queue, forcing sequential access? > >> > >> The scheduling inefficiency you are seeing may be coming from the fact that the > >> block layer does a direct issue of requests, bypassing the elevator, under some > >> conditions. One of these is sync requests on a multiqueue device. We hit this > >> problem on Zoned disks which can easily return an error for write requests > >> without the elevator throttling writes per zones (zone write locking). This > >> problem was discovered by Hans (on CC). > >> > >> I discussed this with Hannes yesterday and we think we have a fix, but we'll > >> need to do a lot of testing as all block devices are potentially impacted by the > >> change, including stacked drivers (DM). Performance regression is scary with any > >> change in that area (see blk_mq_make_request() and use of > >> blk_mq_try_issue_directly() vs blk_mq_sched_insert_request()). > > > > Not sure this one is same with yours, for USB, mq-deadline is used at > > default, and direct issue won't be possible. Direct issue is only used > > in case of none or underlying queues of DM multipath. > > For a multi-queue zoned disk, mq-deadline is also set, but we have observed > unaligned write IO errors for sync writes because of mq-deadline being bypassed > and as a result zones not being write-locked. > > In blk_mq_make_request(), at the end, you have: > > } else if ((q->nr_hw_queues > 1 && is_sync) || (!q->elevator && > !data.hctx->dispatch_busy)) { > blk_mq_try_issue_directly(data.hctx, rq, &cookie); > } else { > blk_mq_sched_insert_request(rq, false, true, true); > } > > Which I read as "for a sync req on a multi-queue device, direct issue", > regardless of the elevator being none or something else. Yeah, looks elevator is bypassed in the above case, which seems a bug. USB storage has only single queue. > > The correct test should probably be: > > } else if (!q->elevator && > ((q->nr_hw_queues > 1 && is_sync) || > !data.hctx->dispatch_busy))) { > blk_mq_try_issue_directly(data.hctx, rq, &cookie); > } else { > blk_mq_sched_insert_request(rq, false, true, true); > } > > That is, never bypass the elevator if one is set. Thoughts ? IMO, elevator shouldn't be bypassed any time, looks it is bypassed in the following branch too, but may not be reached for zone device. blk_mq_make_request() ... } else if (plug && !blk_queue_nomerges(q)) { ... if (same_queue_rq) { data.hctx = same_queue_rq->mq_hctx; trace_block_unplug(q, 1, true); blk_mq_try_issue_directly(data.hctx, same_queue_rq, &cookie); } } Thanks, Ming