On 2020/09/15 10:10, Damien Le Moal wrote: > On 2020/09/15 0:04, Mike Snitzer wrote: >> On Sun, Sep 13 2020 at 8:46pm -0400, >> Damien Le Moal <Damien.LeMoal@xxxxxxx> wrote: >> >>> On 2020/09/12 6:53, Mike Snitzer wrote: >>>> blk_queue_get_max_sectors() has been trained for REQ_OP_WRITE_SAME and >>>> REQ_OP_WRITE_ZEROES yet blk_rq_get_max_sectors() didn't call it for >>>> those operations. >>>> >>>> Also, there is no need to avoid blk_max_size_offset() if >>>> 'chunk_sectors' isn't set because it falls back to 'max_sectors'. >>>> >>>> Signed-off-by: Mike Snitzer <snitzer@xxxxxxxxxx> >>>> --- >>>> include/linux/blkdev.h | 19 +++++++++++++------ >>>> 1 file changed, 13 insertions(+), 6 deletions(-) >>>> >>>> diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h >>>> index bb5636cc17b9..453a3d735d66 100644 >>>> --- a/include/linux/blkdev.h >>>> +++ b/include/linux/blkdev.h >>>> @@ -1070,17 +1070,24 @@ static inline unsigned int blk_rq_get_max_sectors(struct request *rq, >>>> sector_t offset) >>>> { >>>> struct request_queue *q = rq->q; >>>> + int op; >>>> + unsigned int max_sectors; >>>> >>>> if (blk_rq_is_passthrough(rq)) >>>> return q->limits.max_hw_sectors; >>>> >>>> - if (!q->limits.chunk_sectors || >>>> - req_op(rq) == REQ_OP_DISCARD || >>>> - req_op(rq) == REQ_OP_SECURE_ERASE) >>>> - return blk_queue_get_max_sectors(q, req_op(rq)); >>>> + op = req_op(rq); >>>> + max_sectors = blk_queue_get_max_sectors(q, op); >>>> >>>> - return min(blk_max_size_offset(q, offset), >>>> - blk_queue_get_max_sectors(q, req_op(rq))); >>>> + switch (op) { >>>> + case REQ_OP_DISCARD: >>>> + case REQ_OP_SECURE_ERASE: >>>> + case REQ_OP_WRITE_SAME: >>>> + case REQ_OP_WRITE_ZEROES: >>>> + return max_sectors; >>>> + } >>> >>> Doesn't this break md devices ? (I think does use chunk_sectors for stride size, >>> no ?) >>> >>> As mentioned in my reply to Ming's email, this will allow these commands to >>> potentially cross over zone boundaries on zoned block devices, which would be an >>> immediate command failure. >> >> Depending on the implementation it is beneficial to get a large >> discard (one not constrained by chunk_sectors, e.g. dm-stripe.c's >> optimization for handling large discards and issuing N discards, one per >> stripe). Same could apply for other commands. >> >> Like all devices, zoned devices should impose command specific limits in >> the queue_limits (and not lean on chunk_sectors to do a >> one-size-fits-all). > > Yes, understood. But I think that in the case of md, chunk_sectors is used to > indicate the boundary between drives for a raid volume. So it does indeed make > sense to limit the IO size on submission since otherwise, the md driver itself > would have to split that bio again anyway. > >> But that aside, yes I agree I didn't pay close enough attention to the >> implications of deferring the splitting of these commands until they >> were issued to underlying storage. This chunk_sectors early splitting >> override is a bit of a mess... not quite following the logic given we >> were supposed to be waiting to split bios as late as possible. > > My view is that the multipage bvec (BIOs almost as large as we want) and late > splitting is beneficial to get larger effective BIO sent to the device as having > more pages on hand allows bigger segments in the bio instead of always having at > most PAGE_SIZE per segment. The effect of this is very visible with blktrace. A > lot of requests end up being much larger than the device max_segments * page_size. > > However, if there is already a known limit on the BIO size when the BIO is being > built, it does not make much sense to try to grow a bio beyond that limit since > it will have to be split by the driver anyway. chunk_sectors is one such limit > used for md (I think) to indicate boundaries between drives of a raid volume. > And we reuse it (abuse it ?) for zoned block devices to ensure that any command > does not cross over zone boundaries since that triggers errors for writes within > sequential zones or read/write crossing over zones of different types > (conventional->sequential zone boundary). > > I may not have the entire picture correctly here, but so far, this is my > understanding. And I was wrong :) In light of Ming's comment + a little code refresher reading, indeed, chunk_sectors will split BIOs so that *requests* do not exceed that limit, but the initial BIO submission may be much larger regardless of chunk_sectors. Ming, I think the point here is that building a large BIO first and splitting it later (as opposed to limiting the bio size by stopping bio_add_page()) is more efficient as there is only one bio submit instead of many, right ? -- Damien Le Moal Western Digital Research -- dm-devel mailing list dm-devel@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/dm-devel