On 8/5/19 2:08 PM, Damien Le Moal wrote: > On 2019/08/06 5:31, Jens Axboe wrote: >> On 8/5/19 11:31 AM, Jens Axboe wrote: >>> On 8/5/19 11:15 AM, Darrick J. Wong wrote: >>>> Hi Damien, >>>> >>>> I noticed a regression in xfs/747 (an unreleased xfstest for the >>>> xfs_scrub media scanning feature) on 5.3-rc3. I'll condense that down >>>> to a simpler reproducer: >>>> >>>> # dmsetup table >>>> error-test: 0 209 linear 8:48 0 >>>> error-test: 209 1 error >>>> error-test: 210 6446894 linear 8:48 210 >>>> >>>> Basically we have a ~3G /dev/sdd and we set up device mapper to fail IO >>>> for sector 209 and to pass the io to the scsi device everywhere else. >>>> >>>> On 5.3-rc3, performing a directio pread of this range with a < 1M buffer >>>> (in other words, a request for fewer than MAX_BIO_PAGES bytes) yields >>>> EIO like you'd expect: >>>> >>>> # strace -e pread64 xfs_io -d -c 'pread -b 1024k 0k 1120k' /dev/mapper/error-test >>>> pread64(3, 0x7f880e1c7000, 1048576, 0) = -1 EIO (Input/output error) >>>> pread: Input/output error >>>> +++ exited with 0 +++ >>>> >>>> But doing it with a larger buffer succeeds(!): >>>> >>>> # strace -e pread64 xfs_io -d -c 'pread -b 2048k 0k 1120k' /dev/mapper/error-test >>>> pread64(3, "XFSB\0\0\20\0\0\0\0\0\0\fL\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 1146880, 0) = 1146880 >>>> read 1146880/1146880 bytes at offset 0 >>>> 1 MiB, 1 ops; 0.0009 sec (1.124 GiB/sec and 1052.6316 ops/sec) >>>> +++ exited with 0 +++ >>>> >>>> (Note that the part of the buffer corresponding to the dm-error area is >>>> uninitialized) >>>> >>>> On 5.3-rc2, both commands would fail with EIO like you'd expect. The >>>> only change between rc2 and rc3 is commit 0eb6ddfb865c ("block: Fix >>>> __blkdev_direct_IO() for bio fragments"). >>>> >>>> AFAICT we end up in __blkdev_direct_IO with a 1120K buffer, which gets >>>> split into two bios: one for the first BIO_MAX_PAGES worth of data (1MB) >>>> and a second one for the 96k after that. >>>> >>>> I think the problem is that every time we submit a bio, we increase ret >>>> by the size of that bio, but at the time we do that we have no idea if >>>> the bio is going to succeed or not. At the end of the function we do: >>>> >>>> if (!ret) >>>> ret = blk_status_to_errno(dio->bio.bi_status); >>>> >>>> Which means that we only pick up the IO error if we haven't already set >>>> ret. I suppose that was useful for being able to return a short read, >>>> but now that we always increment ret by the size of the bio, we act like >>>> the whole buffer was read. I tried a -rc2 kernel and found that 40% of >>>> the time I'd get an EIO and the rest of the time I got a short read. >>>> >>>> Not sure where to go from here, but something's not right... >>> >>> I'll take a look. >> >> How about this? The old code did: >> >> if (!ret) >> ret = blk_status_to_errno(dio->bio.bi_status); >> if (likely(!ret)) >> ret = dio->size; >> >> where 'ret' was just tracking the error. With 'ret' now being the >> positive IO size, we should overwrite it if ret is >= 0, not just if >> it's zero. >> >> Also kill a use-after-free. >> >> diff --git a/fs/block_dev.c b/fs/block_dev.c >> index a6f7c892cb4a..67c8e87c9481 100644 >> --- a/fs/block_dev.c >> +++ b/fs/block_dev.c >> @@ -386,6 +386,7 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) >> >> ret = 0; >> for (;;) { >> + ssize_t this_size; >> int err; >> >> bio_set_dev(bio, bdev); >> @@ -433,13 +434,14 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) >> polled = true; >> } >> >> + this_size = bio->bi_iter.bi_size; >> qc = submit_bio(bio); >> if (qc == BLK_QC_T_EAGAIN) { >> if (!ret) >> ret = -EAGAIN; >> goto error; >> } >> - ret = dio->size; >> + ret += this_size; >> >> if (polled) >> WRITE_ONCE(iocb->ki_cookie, qc); >> @@ -460,13 +462,14 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) >> atomic_inc(&dio->ref); >> } >> >> + this_size = bio->bi_iter.bi_size; >> qc = submit_bio(bio); >> if (qc == BLK_QC_T_EAGAIN) { >> if (!ret) >> ret = -EAGAIN; >> goto error; >> } >> - ret = dio->size; >> + ret += this_size; >> >> bio = bio_alloc(gfp, nr_pages); >> if (!bio) { >> @@ -494,7 +497,7 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) >> __set_current_state(TASK_RUNNING); >> >> out: >> - if (!ret) >> + if (ret >= 0) >> ret = blk_status_to_errno(dio->bio.bi_status); >> >> bio_put(&dio->bio); >> > > Jens, > > I would set "this_size" when dio->size is being incremented though, to avoid > repeating it. > > if (nowait) > bio->bi_opf |= (REQ_NOWAIT | REQ_NOWAIT_INLINE); > > + this_size = bio->bi_iter.bi_size; > - dio->size += bio->bi_iter.bi_size; > + dio->size += this_size; > pos += bio->bi_iter.bi_size; > > In any case, looking again at this code, it looks like there is a > problem with dio->size being incremented early, even for fragments > that get BLK_QC_T_EAGAIN, because dio->size is being used in > blkdev_bio_end_io(). So an incorrect size can be reported to user > space in that case on completion (e.g. large asynchronous no-wait dio > that cannot be issued in one go). > > So maybe something like this ? (completely untested) I think that looks pretty good, I like not double accounting with this_size and dio->size, and we retain the old style ordering for the ret value. -- Jens Axboe