Re: [PATCH V4 13/45] block: blk-merge: try to make front segments in full size

From: Dmitry Osipenko
Date: Wed Jan 10 2018 - 10:41:27 EST


On 10.01.2018 05:40, Ming Lei wrote:
> On Tue, Jan 09, 2018 at 08:02:53PM +0300, Dmitry Osipenko wrote:
>> On 09.01.2018 17:33, Ming Lei wrote:
>>> On Tue, Jan 09, 2018 at 04:18:39PM +0300, Dmitry Osipenko wrote:
>>>> On 09.01.2018 05:34, Ming Lei wrote:
>>>>> On Tue, Jan 09, 2018 at 12:09:27AM +0300, Dmitry Osipenko wrote:
>>>>>> On 18.12.2017 15:22, Ming Lei wrote:
>>>>>>> When merging one bvec into segment, if the bvec is too big
>>>>>>> to merge, current policy is to move the whole bvec into another
>>>>>>> new segment.
>>>>>>>
>>>>>>> This patchset changes the policy into trying to maximize size of
>>>>>>> front segments, that means in above situation, part of bvec
>>>>>>> is merged into current segment, and the remainder is put
>>>>>>> into next segment.
>>>>>>>
>>>>>>> This patch prepares for support multipage bvec because
>>>>>>> it can be quite common to see this case and we should try
>>>>>>> to make front segments in full size.
>>>>>>>
>>>>>>> Signed-off-by: Ming Lei <ming.lei@xxxxxxxxxx>
>>>>>>> ---
>>>>>>> block/blk-merge.c | 54 +++++++++++++++++++++++++++++++++++++++++++++++++-----
>>>>>>> 1 file changed, 49 insertions(+), 5 deletions(-)
>>>>>>>
>>>>>>> diff --git a/block/blk-merge.c b/block/blk-merge.c
>>>>>>> index a476337a8ff4..42ceb89bc566 100644
>>>>>>> --- a/block/blk-merge.c
>>>>>>> +++ b/block/blk-merge.c
>>>>>>> @@ -109,6 +109,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>>>>>> bool do_split = true;
>>>>>>> struct bio *new = NULL;
>>>>>>> const unsigned max_sectors = get_max_io_size(q, bio);
>>>>>>> + unsigned advance = 0;
>>>>>>>
>>>>>>> bio_for_each_segment(bv, bio, iter) {
>>>>>>> /*
>>>>>>> @@ -134,12 +135,32 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>>>>>> }
>>>>>>>
>>>>>>> if (bvprvp && blk_queue_cluster(q)) {
>>>>>>> - if (seg_size + bv.bv_len > queue_max_segment_size(q))
>>>>>>> - goto new_segment;
>>>>>>> if (!BIOVEC_PHYS_MERGEABLE(bvprvp, &bv))
>>>>>>> goto new_segment;
>>>>>>> if (!BIOVEC_SEG_BOUNDARY(q, bvprvp, &bv))
>>>>>>> goto new_segment;
>>>>>>> + if (seg_size + bv.bv_len > queue_max_segment_size(q)) {
>>>>>>> + /*
>>>>>>> + * On assumption is that initial value of
>>>>>>> + * @seg_size(equals to bv.bv_len) won't be
>>>>>>> + * bigger than max segment size, but will
>>>>>>> + * becomes false after multipage bvec comes.
>>>>>>> + */
>>>>>>> + advance = queue_max_segment_size(q) - seg_size;
>>>>>>> +
>>>>>>> + if (advance > 0) {
>>>>>>> + seg_size += advance;
>>>>>>> + sectors += advance >> 9;
>>>>>>> + bv.bv_len -= advance;
>>>>>>> + bv.bv_offset += advance;
>>>>>>> + }
>>>>>>> +
>>>>>>> + /*
>>>>>>> + * Still need to put remainder of current
>>>>>>> + * bvec into a new segment.
>>>>>>> + */
>>>>>>> + goto new_segment;
>>>>>>> + }
>>>>>>>
>>>>>>> seg_size += bv.bv_len;
>>>>>>> bvprv = bv;
>>>>>>> @@ -161,6 +182,12 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>>>>>> seg_size = bv.bv_len;
>>>>>>> sectors += bv.bv_len >> 9;
>>>>>>>
>>>>>>> + /* restore the bvec for iterator */
>>>>>>> + if (advance) {
>>>>>>> + bv.bv_len += advance;
>>>>>>> + bv.bv_offset -= advance;
>>>>>>> + advance = 0;
>>>>>>> + }
>>>>>>> }
>>>>>>>
>>>>>>> do_split = false;
>>>>>>> @@ -361,16 +388,29 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>>>>>> {
>>>>>>>
>>>>>>> int nbytes = bvec->bv_len;
>>>>>>> + unsigned advance = 0;
>>>>>>>
>>>>>>> if (*sg && *cluster) {
>>>>>>> - if ((*sg)->length + nbytes > queue_max_segment_size(q))
>>>>>>> - goto new_segment;
>>>>>>> -
>>>>>>> if (!BIOVEC_PHYS_MERGEABLE(bvprv, bvec))
>>>>>>> goto new_segment;
>>>>>>> if (!BIOVEC_SEG_BOUNDARY(q, bvprv, bvec))
>>>>>>> goto new_segment;
>>>>>>>
>>>>>>> + /*
>>>>>>> + * try best to merge part of the bvec into previous
>>>>>>> + * segment and follow same policy with
>>>>>>> + * blk_bio_segment_split()
>>>>>>> + */
>>>>>>> + if ((*sg)->length + nbytes > queue_max_segment_size(q)) {
>>>>>>> + advance = queue_max_segment_size(q) - (*sg)->length;
>>>>>>> + if (advance) {
>>>>>>> + (*sg)->length += advance;
>>>>>>> + bvec->bv_offset += advance;
>>>>>>> + bvec->bv_len -= advance;
>>>>>>> + }
>>>>>>> + goto new_segment;
>>>>>>> + }
>>>>>>> +
>>>>>>> (*sg)->length += nbytes;
>>>>>>> } else {
>>>>>>> new_segment:
>>>>>>> @@ -393,6 +433,10 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>>>>>>
>>>>>>> sg_set_page(*sg, bvec->bv_page, nbytes, bvec->bv_offset);
>>>>>>> (*nsegs)++;
>>>>>>> +
>>>>>>> + /* for making iterator happy */
>>>>>>> + bvec->bv_offset -= advance;
>>>>>>> + bvec->bv_len += advance;
>>>>>>> }
>>>>>>> *bvprv = *bvec;
>>>>>>> }
>>>>>>>
>>>>>>
>>>>>> Hello,
>>>>>>
>>>>>> This patch breaks MMC on next-20180108, in particular MMC doesn't work anymore
>>>>>> with this patch on NVIDIA Tegra20:
>>>>>>
>>>>>> <3>[ 36.622253] print_req_error: I/O error, dev mmcblk1, sector 512
>>>>>> <3>[ 36.671233] print_req_error: I/O error, dev mmcblk2, sector 128
>>>>>> <3>[ 36.711308] print_req_error: I/O error, dev mmcblk1, sector 31325304
>>>>>> <3>[ 36.749232] print_req_error: I/O error, dev mmcblk2, sector 512
>>>>>> <3>[ 36.761235] print_req_error: I/O error, dev mmcblk1, sector 31325816
>>>>>> <3>[ 36.832039] print_req_error: I/O error, dev mmcblk2, sector 31259768
>>>>>> <3>[ 99.793248] print_req_error: I/O error, dev mmcblk1, sector 31323136
>>>>>> <3>[ 99.982043] print_req_error: I/O error, dev mmcblk1, sector 929792
>>>>>> <3>[ 99.986301] print_req_error: I/O error, dev mmcblk1, sector 930816
>>>>>> <3>[ 100.293624] print_req_error: I/O error, dev mmcblk1, sector 932864
>>>>>> <3>[ 100.466839] print_req_error: I/O error, dev mmcblk1, sector 947200
>>>>>> <3>[ 100.642955] print_req_error: I/O error, dev mmcblk1, sector 949248
>>>>>> <3>[ 100.818838] print_req_error: I/O error, dev mmcblk1, sector 230400
>>>>>>
>>>>>> Any attempt of mounting MMC block dev ends with a kernel crash. Reverting this
>>>>>> patch fixes the issue.
>>>>>
>>>>> Hi Dmitry,
>>>>>
>>>>> Thanks for your report!
>>>>>
>>>>> Could you share us what the segment limits are on your MMC?
>>>>>
>>>>> cat /sys/block/mmcN/queue/max_segment_size
>>>>> cat /sys/block/mmcN/queue/max_segments
>>>>>
>>>>> Please test the following patch to see if your issue can be fixed?
>>>>>
>>>>> ---
>>>>> diff --git a/block/blk-merge.c b/block/blk-merge.c
>>>>> index 446f63e076aa..cfab36c26608 100644
>>>>> --- a/block/blk-merge.c
>>>>> +++ b/block/blk-merge.c
>>>>> @@ -431,12 +431,14 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>>>>
>>>>> sg_set_page(*sg, bvec->bv_page, nbytes, bvec->bv_offset);
>>>>> (*nsegs)++;
>>>>> + }
>>>>>
>>>>> + *bvprv = *bvec;
>>>>> + if (advance) {
>>>>> /* for making iterator happy */
>>>>> bvec->bv_offset -= advance;
>>>>> bvec->bv_len += advance;
>>>>> }
>>>>> - *bvprv = *bvec;
>>>>> }
>>>>>
>>>>> static inline int __blk_bvec_map_sg(struct request_queue *q, struct bio_vec bv,
>>>>
>>>> Hi Ming,
>>>>
>>>> I've tried your patch and unfortunately it doesn't help with the issue.
>>>>
>>>> Here are the segment limits:
>>>>
>>>> # cat /sys/block/mmc*/queue/max_segment_size
>>>> 65535
>>>
>>> Hi Dmitry,
>>>
>>> The 'max_segment_size' of 65535 should be the reason, could you test the
>>> following patch?
>>>
>>> ---
>>> diff --git a/block/blk-merge.c b/block/blk-merge.c
>>> index 446f63e076aa..38a66e3e678e 100644
>>> --- a/block/blk-merge.c
>>> +++ b/block/blk-merge.c
>>> @@ -12,6 +12,8 @@
>>>
>>> #include "blk.h"
>>>
>>> +#define sector_align(x) ALIGN_DOWN(x, 512)
>>> +
>>> static struct bio *blk_bio_discard_split(struct request_queue *q,
>>> struct bio *bio,
>>> struct bio_set *bs,
>>> @@ -109,7 +111,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>> bool do_split = true;
>>> struct bio *new = NULL;
>>> const unsigned max_sectors = get_max_io_size(q, bio);
>>> - unsigned advance = 0;
>>> + int advance = 0;
>>>
>>> bio_for_each_segment(bv, bio, iter) {
>>> /*
>>> @@ -144,8 +146,9 @@ static struct bio *blk_bio_segment_split(struct request_queue *q,
>>> * bigger than max segment size, but this
>>> * becomes false after multipage bvecs.
>>> */
>>> - advance = queue_max_segment_size(q) - seg_size;
>>> -
>>> + advance = sector_align(
>>> + queue_max_segment_size(q) -
>>> + seg_size);
>>> if (advance > 0) {
>>> seg_size += advance;
>>> sectors += advance >> 9;
>>> @@ -386,7 +389,7 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>> {
>>>
>>> int nbytes = bvec->bv_len;
>>> - unsigned advance = 0;
>>> + int advance = 0;
>>>
>>> if (*sg && *cluster) {
>>> if (!BIOVEC_PHYS_MERGEABLE(bvprv, bvec))
>>> @@ -400,8 +403,9 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>> * blk_bio_segment_split()
>>> */
>>> if ((*sg)->length + nbytes > queue_max_segment_size(q)) {
>>> - advance = queue_max_segment_size(q) - (*sg)->length;
>>> - if (advance) {
>>> + advance = sector_align(queue_max_segment_size(q) -
>>> + (*sg)->length);
>>> + if (advance > 0) {
>>> (*sg)->length += advance;
>>> bvec->bv_offset += advance;
>>> bvec->bv_len -= advance;
>>> @@ -431,12 +435,14 @@ __blk_segment_map_sg(struct request_queue *q, struct bio_vec *bvec,
>>>
>>> sg_set_page(*sg, bvec->bv_page, nbytes, bvec->bv_offset);
>>> (*nsegs)++;
>>> + }
>>>
>>> + *bvprv = *bvec;
>>> + if (advance > 0) {
>>> /* for making iterator happy */
>>> bvec->bv_offset -= advance;
>>> bvec->bv_len += advance;
>>> }
>>> - *bvprv = *bvec;
>>> }
>>>
>>> static inline int __blk_bvec_map_sg(struct request_queue *q, struct bio_vec bv,
>>
>> This patch doesn't help either.
>
> OK, I will send a revert later.
>
> Thinking of the patch further, we don't need this kind of logic for
> multipage bvec at all since almost all bvecs won't be contiguous if
> bio_add_page() is used after multipage bvec is enabled.

Revert works for me, thanks.