On 2019/08/06 5:31, Jens Axboe wrote: > On 8/5/19 11:31 AM, Jens Axboe wrote: >> On 8/5/19 11:15 AM, Darrick J. Wong wrote: >>> Hi Damien, >>> >>> I noticed a regression in xfs/747 (an unreleased xfstest for the >>> xfs_scrub media scanning feature) on 5.3-rc3. I'll condense that down >>> to a simpler reproducer: >>> >>> # dmsetup table >>> error-test: 0 209 linear 8:48 0 >>> error-test: 209 1 error >>> error-test: 210 6446894 linear 8:48 210 >>> >>> Basically we have a ~3G /dev/sdd and we set up device mapper to fail IO >>> for sector 209 and to pass the io to the scsi device everywhere else. >>> >>> On 5.3-rc3, performing a directio pread of this range with a < 1M buffer >>> (in other words, a request for fewer than MAX_BIO_PAGES bytes) yields >>> EIO like you'd expect: >>> >>> # strace -e pread64 xfs_io -d -c 'pread -b 1024k 0k 1120k' /dev/mapper/error-test >>> pread64(3, 0x7f880e1c7000, 1048576, 0) = -1 EIO (Input/output error) >>> pread: Input/output error >>> +++ exited with 0 +++ >>> >>> But doing it with a larger buffer succeeds(!): >>> >>> # strace -e pread64 xfs_io -d -c 'pread -b 2048k 0k 1120k' /dev/mapper/error-test >>> pread64(3, "XFSB\0\0\20\0\0\0\0\0\0\fL\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0"..., 1146880, 0) = 1146880 >>> read 1146880/1146880 bytes at offset 0 >>> 1 MiB, 1 ops; 0.0009 sec (1.124 GiB/sec and 1052.6316 ops/sec) >>> +++ exited with 0 +++ >>> >>> (Note that the part of the buffer corresponding to the dm-error area is >>> uninitialized) >>> >>> On 5.3-rc2, both commands would fail with EIO like you'd expect. The >>> only change between rc2 and rc3 is commit 0eb6ddfb865c ("block: Fix >>> __blkdev_direct_IO() for bio fragments"). >>> >>> AFAICT we end up in __blkdev_direct_IO with a 1120K buffer, which gets >>> split into two bios: one for the first BIO_MAX_PAGES worth of data (1MB) >>> and a second one for the 96k after that. >>> >>> I think the problem is that every time we submit a bio, we increase ret >>> by the size of that bio, but at the time we do that we have no idea if >>> the bio is going to succeed or not. At the end of the function we do: >>> >>> if (!ret) >>> ret = blk_status_to_errno(dio->bio.bi_status); >>> >>> Which means that we only pick up the IO error if we haven't already set >>> ret. I suppose that was useful for being able to return a short read, >>> but now that we always increment ret by the size of the bio, we act like >>> the whole buffer was read. I tried a -rc2 kernel and found that 40% of >>> the time I'd get an EIO and the rest of the time I got a short read. >>> >>> Not sure where to go from here, but something's not right... >> >> I'll take a look. > > How about this? The old code did: > > if (!ret) > ret = blk_status_to_errno(dio->bio.bi_status); > if (likely(!ret)) > ret = dio->size; > > where 'ret' was just tracking the error. With 'ret' now being the > positive IO size, we should overwrite it if ret is >= 0, not just if > it's zero. > > Also kill a use-after-free. > > diff --git a/fs/block_dev.c b/fs/block_dev.c > index a6f7c892cb4a..67c8e87c9481 100644 > --- a/fs/block_dev.c > +++ b/fs/block_dev.c > @@ -386,6 +386,7 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) > > ret = 0; > for (;;) { > + ssize_t this_size; > int err; > > bio_set_dev(bio, bdev); > @@ -433,13 +434,14 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) > polled = true; > } > > + this_size = bio->bi_iter.bi_size; > qc = submit_bio(bio); > if (qc == BLK_QC_T_EAGAIN) { > if (!ret) > ret = -EAGAIN; > goto error; > } > - ret = dio->size; > + ret += this_size; > > if (polled) > WRITE_ONCE(iocb->ki_cookie, qc); > @@ -460,13 +462,14 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) > atomic_inc(&dio->ref); > } > > + this_size = bio->bi_iter.bi_size; > qc = submit_bio(bio); > if (qc == BLK_QC_T_EAGAIN) { > if (!ret) > ret = -EAGAIN; > goto error; > } > - ret = dio->size; > + ret += this_size; > > bio = bio_alloc(gfp, nr_pages); > if (!bio) { > @@ -494,7 +497,7 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) > __set_current_state(TASK_RUNNING); > > out: > - if (!ret) > + if (ret >= 0) > ret = blk_status_to_errno(dio->bio.bi_status); > > bio_put(&dio->bio); > Jens, I would set "this_size" when dio->size is being incremented though, to avoid repeating it. if (nowait) bio->bi_opf |= (REQ_NOWAIT | REQ_NOWAIT_INLINE); + this_size = bio->bi_iter.bi_size; - dio->size += bio->bi_iter.bi_size; + dio->size += this_size; pos += bio->bi_iter.bi_size; In any case, looking again at this code, it looks like there is a problem with dio->size being incremented early, even for fragments that get BLK_QC_T_EAGAIN, because dio->size is being used in blkdev_bio_end_io(). So an incorrect size can be reported to user space in that case on completion (e.g. large asynchronous no-wait dio that cannot be issued in one go). So maybe something like this ? (completely untested) diff --git a/fs/block_dev.c b/fs/block_dev.c index 75cc7f424b3a..77714e03c21e 100644 --- a/fs/block_dev.c +++ b/fs/block_dev.c @@ -349,7 +349,7 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) loff_t pos = iocb->ki_pos; blk_qc_t qc = BLK_QC_T_NONE; gfp_t gfp; - ssize_t ret; + ssize_t ret = 0; if ((pos | iov_iter_alignment(iter)) & (bdev_logical_block_size(bdev) - 1)) @@ -386,6 +386,7 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) ret = 0; for (;;) { + size_t this_size; int err; bio_set_dev(bio, bdev); @@ -421,7 +422,7 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) if (nowait) bio->bi_opf |= (REQ_NOWAIT | REQ_NOWAIT_INLINE); - dio->size += bio->bi_iter.bi_size; + this_size = bio->bi_iter.bi_size; pos += bio->bi_iter.bi_size; nr_pages = iov_iter_npages(iter, BIO_MAX_PAGES); @@ -435,11 +436,11 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) qc = submit_bio(bio); if (qc == BLK_QC_T_EAGAIN) { - if (!ret) + if (!dio->size) ret = -EAGAIN; goto error; } - ret = dio->size; + dio->size += this_size; if (polled) WRITE_ONCE(iocb->ki_cookie, qc); @@ -462,15 +463,15 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) qc = submit_bio(bio); if (qc == BLK_QC_T_EAGAIN) { - if (!ret) + if (!dio->size) ret = -EAGAIN; goto error; } - ret = dio->size; + dio->size += this_size; bio = bio_alloc(gfp, nr_pages); if (!bio) { - if (!ret) + if (!dio->size) ret = -EAGAIN; goto error; } @@ -496,10 +497,15 @@ __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) out: if (!ret) ret = blk_status_to_errno(dio->bio.bi_status); + if (likely(!ret)) + ret = dio->size; bio_put(&dio->bio); return ret; -- Damien Le Moal Western Digital Research