Re: [PATCH V4 13/45] block: blk-merge: try to make front segments in full size

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 09.01.2018 17:33, Ming Lei wrote:
> On Tue, Jan 09, 2018 at 04:18:39PM +0300, Dmitry Osipenko wrote:
>> On 09.01.2018 05:34, Ming Lei wrote:
>>> On Tue, Jan 09, 2018 at 12:09:27AM +0300, Dmitry Osipenko wrote:
>>>> On 18.12.2017 15:22, Ming Lei wrote:
>>>>> When merging one bvec into segment, if the bvec is too big
>>>>> to merge, current policy is to move the whole bvec into another
>>>>> new segment.
>>>>>
>>>>> This patchset changes the policy into trying to maximize size of
>>>>> front segments, that means in above situation, part of bvec
>>>>> is merged into current segment, and the remainder is put
>>>>> into next segment.
>>>>>
>>>>> This patch prepares for support multipage bvec because
>>>>> it can be quite common to see this case and we should try
>>>>> to make front segments in full size.
>>>>>
>>>>> Signed-off-by: Ming Lei <ming.lei@xxxxxxxxxx>
>>>>> ---
>>>>>  block/blk-merge.c | 54 +++++++++++++++++++++++++++++++++++++++++++++++++-----
>>>>>  1 file changed, 49 insertions(+), 5 deletions(-)
>>>>>
>>>>> diff --git a/block/blk-merge.c b/block/blk-merge.c
>>>>> index a476337a8ff4..42ceb89bc566 100644
>>>>> --- a/block/blk-merge.c
>>>>> +++ b/block/blk-merge.c
>>>>> @@ -109,6 +109,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>>>>  	bool do_split = true;
>>>>>  	struct bio *new = NULL;
>>>>>  	const unsigned max_sectors = get_max_io_size(q, bio);
>>>>> +	unsigned advance = 0;
>>>>>  
>>>>>  	bio_for_each_segment(bv, bio, iter) {
>>>>>  		/*
>>>>> @@ -134,12 +135,32 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>>>>  		}
>>>>>  
>>>>>  		if (bvprvp && blk_queue_cluster(q)) {
>>>>> -			if (seg_size + bv.bv_len > queue_max_segment_size(q))
>>>>> -				goto new_segment;
>>>>>  			if (!BIOVEC_PHYS_MERGEABLE(bvprvp, &bv))
>>>>>  				goto new_segment;
>>>>>  			if (!BIOVEC_SEG_BOUNDARY(q, bvprvp, &bv))
>>>>>  				goto new_segment;
>>>>> +			if (seg_size + bv.bv_len > queue_max_segment_size(q)) {
>>>>> +				/*
>>>>> +				 * On assumption is that initial value of
>>>>> +				 * @seg_size(equals to bv.bv_len) won't be
>>>>> +				 * bigger than max segment size, but will
>>>>> +				 * becomes false after multipage bvec comes.
>>>>> +				 */
>>>>> +				advance = queue_max_segment_size(q) - seg_size;
>>>>> +
>>>>> +				if (advance > 0) {
>>>>> +					seg_size += advance;
>>>>> +					sectors += advance >> 9;
>>>>> +					bv.bv_len -= advance;
>>>>> +					bv.bv_offset += advance;
>>>>> +				}
>>>>> +
>>>>> +				/*
>>>>> +				 * Still need to put remainder of current
>>>>> +				 * bvec into a new segment.
>>>>> +				 */
>>>>> +				goto new_segment;
>>>>> +			}
>>>>>  
>>>>>  			seg_size += bv.bv_len;
>>>>>  			bvprv = bv;
>>>>> @@ -161,6 +182,12 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>>>>  		seg_size = bv.bv_len;
>>>>>  		sectors += bv.bv_len >> 9;
>>>>>  
>>>>> +		/* restore the bvec for iterator */
>>>>> +		if (advance) {
>>>>> +			bv.bv_len += advance;
>>>>> +			bv.bv_offset -= advance;
>>>>> +			advance = 0;
>>>>> +		}
>>>>>  	}
>>>>>  
>>>>>  	do_split = false;
>>>>> @@ -361,16 +388,29 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>>>>  {
>>>>>  
>>>>>  	int nbytes = bvec->bv_len;
>>>>> +	unsigned advance = 0;
>>>>>  
>>>>>  	if (*sg && *cluster) {
>>>>> -		if ((*sg)->length + nbytes > queue_max_segment_size(q))
>>>>> -			goto new_segment;
>>>>> -
>>>>>  		if (!BIOVEC_PHYS_MERGEABLE(bvprv, bvec))
>>>>>  			goto new_segment;
>>>>>  		if (!BIOVEC_SEG_BOUNDARY(q, bvprv, bvec))
>>>>>  			goto new_segment;
>>>>>  
>>>>> +		/*
>>>>> +		 * try best to merge part of the bvec into previous
>>>>> +		 * segment and follow same policy with
>>>>> +		 * blk_bio_segment_split()
>>>>> +		 */
>>>>> +		if ((*sg)->length + nbytes > queue_max_segment_size(q)) {
>>>>> +			advance = queue_max_segment_size(q) - (*sg)->length;
>>>>> +			if (advance) {
>>>>> +				(*sg)->length += advance;
>>>>> +				bvec->bv_offset += advance;
>>>>> +				bvec->bv_len -= advance;
>>>>> +			}
>>>>> +			goto new_segment;
>>>>> +		}
>>>>> +
>>>>>  		(*sg)->length += nbytes;
>>>>>  	} else {
>>>>>  new_segment:
>>>>> @@ -393,6 +433,10 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>>>>  
>>>>>  		sg_set_page(*sg, bvec->bv_page, nbytes, bvec->bv_offset);
>>>>>  		(*nsegs)++;
>>>>> +
>>>>> +		/* for making iterator happy */
>>>>> +		bvec->bv_offset -= advance;
>>>>> +		bvec->bv_len += advance;
>>>>>  	}
>>>>>  	*bvprv = *bvec;
>>>>>  }
>>>>>
>>>>
>>>> Hello,
>>>>
>>>> This patch breaks MMC on next-20180108, in particular MMC doesn't work anymore
>>>> with this patch on NVIDIA Tegra20:
>>>>
>>>> <3>[   36.622253] print_req_error: I/O error, dev mmcblk1, sector 512
>>>> <3>[   36.671233] print_req_error: I/O error, dev mmcblk2, sector 128
>>>> <3>[   36.711308] print_req_error: I/O error, dev mmcblk1, sector 31325304
>>>> <3>[   36.749232] print_req_error: I/O error, dev mmcblk2, sector 512
>>>> <3>[   36.761235] print_req_error: I/O error, dev mmcblk1, sector 31325816
>>>> <3>[   36.832039] print_req_error: I/O error, dev mmcblk2, sector 31259768
>>>> <3>[   99.793248] print_req_error: I/O error, dev mmcblk1, sector 31323136
>>>> <3>[   99.982043] print_req_error: I/O error, dev mmcblk1, sector 929792
>>>> <3>[   99.986301] print_req_error: I/O error, dev mmcblk1, sector 930816
>>>> <3>[  100.293624] print_req_error: I/O error, dev mmcblk1, sector 932864
>>>> <3>[  100.466839] print_req_error: I/O error, dev mmcblk1, sector 947200
>>>> <3>[  100.642955] print_req_error: I/O error, dev mmcblk1, sector 949248
>>>> <3>[  100.818838] print_req_error: I/O error, dev mmcblk1, sector 230400
>>>>
>>>> Any attempt of mounting MMC block dev ends with a kernel crash. Reverting this
>>>> patch fixes the issue.
>>>
>>> Hi Dmitry,
>>>
>>> Thanks for your report!
>>>
>>> Could you share us what the segment limits are on your MMC?
>>>
>>> 	cat /sys/block/mmcN/queue/max_segment_size
>>> 	cat /sys/block/mmcN/queue/max_segments
>>>
>>> Please test the following patch to see if your issue can be fixed?
>>>
>>> ---
>>> diff --git a/block/blk-merge.c b/block/blk-merge.c
>>> index 446f63e076aa..cfab36c26608 100644
>>> --- a/block/blk-merge.c
>>> +++ b/block/blk-merge.c
>>> @@ -431,12 +431,14 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>>  
>>>  		sg_set_page(*sg, bvec->bv_page, nbytes, bvec->bv_offset);
>>>  		(*nsegs)++;
>>> +	}
>>>  
>>> +	*bvprv = *bvec;
>>> +	if (advance) {
>>>  		/* for making iterator happy */
>>>  		bvec->bv_offset -= advance;
>>>  		bvec->bv_len += advance;
>>>  	}
>>> -	*bvprv = *bvec;
>>>  }
>>>  
>>>  static inline int __blk_bvec_map_sg(struct request_queue *q, struct bio_vec bv,
>>
>> Hi Ming,
>>
>> I've tried your patch and unfortunately it doesn't help with the issue.
>>
>> Here are the segment limits:
>>
>> # cat /sys/block/mmc*/queue/max_segment_size
>> 65535
> 
> Hi Dmitry,
> 
> The 'max_segment_size' of 65535 should be the reason, could you test the
> following patch?
> 
> ---
> diff --git a/block/blk-merge.c b/block/blk-merge.c
> index 446f63e076aa..38a66e3e678e 100644
> --- a/block/blk-merge.c
> +++ b/block/blk-merge.c
> @@ -12,6 +12,8 @@
>  
>  #include "blk.h"
>  
> +#define sector_align(x)   ALIGN_DOWN(x, 512)
> +
>  static struct bio *blk_bio_discard_split(struct request_queue *q,
>  					 struct bio *bio,
>  					 struct bio_set *bs,
> @@ -109,7 +111,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>  	bool do_split = true;
>  	struct bio *new = NULL;
>  	const unsigned max_sectors = get_max_io_size(q, bio);
> -	unsigned advance = 0;
> +	int advance = 0;
>  
>  	bio_for_each_segment(bv, bio, iter) {
>  		/*
> @@ -144,8 +146,9 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>  				 * bigger than max segment size, but this
>  				 * becomes false after multipage bvecs.
>  				 */
> -				advance = queue_max_segment_size(q) - seg_size;
> -
> +				advance = sector_align(
> +						queue_max_segment_size(q) -
> +						seg_size);
>  				if (advance > 0) {
>  					seg_size += advance;
>  					sectors += advance >> 9;
> @@ -386,7 +389,7 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>  {
>  
>  	int nbytes = bvec->bv_len;
> -	unsigned advance = 0;
> +	int advance = 0;
>  
>  	if (*sg && *cluster) {
>  		if (!BIOVEC_PHYS_MERGEABLE(bvprv, bvec))
> @@ -400,8 +403,9 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>  		 * blk_bio_segment_split()
>  		 */
>  		if ((*sg)->length + nbytes > queue_max_segment_size(q)) {
> -			advance = queue_max_segment_size(q) - (*sg)->length;
> -			if (advance) {
> +			advance = sector_align(queue_max_segment_size(q) -
> +					(*sg)->length);
> +			if (advance > 0) {
>  				(*sg)->length += advance;
>  				bvec->bv_offset += advance;
>  				bvec->bv_len -= advance;
> @@ -431,12 +435,14 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>  
>  		sg_set_page(*sg, bvec->bv_page, nbytes, bvec->bv_offset);
>  		(*nsegs)++;
> +	}
>  
> +	*bvprv = *bvec;
> +	if (advance > 0) {
>  		/* for making iterator happy */
>  		bvec->bv_offset -= advance;
>  		bvec->bv_len += advance;
>  	}
> -	*bvprv = *bvec;
>  }
>  
>  static inline int __blk_bvec_map_sg(struct request_queue *q, struct bio_vec bv,

This patch doesn't help either.



[Index of Archives]     [Linux Ext4 Filesystem]     [Union Filesystem]     [Filesystem Testing]     [Ceph Users]     [Ecryptfs]     [AutoFS]     [Kernel Newbies]     [Share Photos]     [Security]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux Cachefs]     [Reiser Filesystem]     [Linux RAID]     [Samba]     [Device Mapper]     [CEPH Development]
  Powered by Linux