On Fri, Mar 29, 2019 at 12:32 AM Johannes Thumshirn <jthumshirn@xxxxxxx> wrote: > > On 25/03/2019 02:04, kernel test robot wrote: > [...] > > nvme/005 (reset local loopback target) > > runtime ... > > nvme/005 (reset local loopback target) [failed] > > runtime ... 0.596s > > something found in dmesg: > > [ 24.160182] run blktests nvme/005 at 2019-03-24 07:35:16 > > [ 24.346189] nvmet: adding nsid 1 to subsystem blktests-subsystem-1 > > [ 24.373089] nvmet: creating controller 1 for subsystem blktests-subsystem-1 for NQN nqn.2014-08.org.nvmexpress:uuid:dbd85962-80d1-4872-ac0f-d0214f3b1131. > > [ 24.394247] WARNING: CPU: 1 PID: 881 at /kbuild/src/x86_64/block/blk-merge.c:608 blk_rq_map_sg+0x5f0/0x6c0 > > [ 24.396082] Modules linked in: nvme_loop nvme_fabrics nvmet nvme_core loop crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel ata_generic sd_mod pata_acpi sg bochs_drm ttm ppdev drm_kms_helper syscopyarea sysfillrect sysimgblt fb_sys_fops snd_pcm drm ata_piix snd_timer aesni_intel crypto_simd snd cryptd glue_helper libata joydev soundcore pcspkr serio_raw virtio_scsi i2c_piix4 parport_pc floppy parport ip_tables > > [ 24.403009] CPU: 1 PID: 881 Comm: nvme Not tainted 5.0.0-rc6-00148-g4e366a7 #1 > > [ 24.404372] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.10.2-1 04/01/2014 > > [ 24.405930] RIP: 0010:blk_rq_map_sg+0x5f0/0x6c0 > > [ 24.406808] Code: 41 8b 91 10 01 00 00 8b 43 24 85 c2 0f 84 e3 fb ff ff f7 d0 21 d0 83 c0 01 41 01 40 0c 01 83 d0 00 00 00 e9 cd fb ff ff 0f 0b <0f> 0b e9 c3 fc ff ff 80 3d 66 f3 1d 01 00 0f 85 88 fb ff ff 48 c7 > > [ 24.410274] RSP: 0000:ffffb95c409d7908 EFLAGS: 00010202 > > ... > > (See '/lkp/benchmarks/blktests/blktests/results/nodev/nvme/005.dmesg' for the entire message) > > OK we're tripping over this in blk_rq_map_sg(): > > 607 /* > 608 * Something must have been wrong if the figured number of > 609 * segment is bigger than number of req's physical segments > 610 */ > 611 WARN_ON(nsegs > blk_rq_nr_phys_segments(rq)); > > In all cases I could reproduce it I had nsegs == 2 and > rq->nr_phys_segments == 1 (rq->bio->bi_phys_segments == 1). > > Adding this into nvme-loop "fixes" the issue but I don't think this is > the correct way to do > > --- a/drivers/nvme/target/loop.c > +++ b/drivers/nvme/target/loop.c > @@ -387,6 +387,7 @@ static int nvme_loop_configure_admin_que > goto out_free_tagset; > } > > + blk_queue_segment_boundary(ctrl->ctrl.admin_q, PAGE_SIZE - 1); The above isn't needed for linus tree, however it is required for 5.0 and older releases. But nvme-loop target have other problems, and I'd suggest you to apply the following patch: https://git.kernel.org/pub/scm/linux/kernel/git/axboe/linux-block.git/commit/?h=for-linus&id=02db99548d3608a625cf481cff2bb7b626829b3f > error = nvmf_connect_admin_queue(&ctrl->ctrl); > if (error) > goto out_cleanup_queue; > > What raises my suspicion is this code fragment from bio_add_pc_page(): > > 731 > 732 /* If we may be able to merge these biovecs, force a recount */ > 733 if (bio->bi_vcnt > 1 && biovec_phys_mergeable(q, bvec - 1, bvec)) > 734 bio->bi_phys_segments = -1; The above change from you might cause issue given some queue's .bi_phys_segments is 255. > > This -1 (from this patch) would fit the one less than 'nsegs' but after > days of searching and staring at the code, I cannot find where we loose > the one. Before this patch set we would only have cleared the > BIO_SEG_VALID flag and thus we never run out of sync. I'd suggest you to address the following comments first before working on your patch: https://lore.kernel.org/linux-block/CACVXFVNPW1TOsCaAibc-YJHDEdK95Y8-v1rHNaab7oV=eb+8TA@xxxxxxxxxxxxxx/ https://lore.kernel.org/linux-block/CACVXFVOq1ngjpD62SreSZda5k_WJjxOZRHwpqt6si7bRv=O46Q@xxxxxxxxxxxxxx/ > > Adding a call to blk_recount_segments() doesn't do any difference here. > > It can of be a read herring as well. > > What makes me suspicious is the fact that I can't reliably trigger it by > running nvme/005 but have to run it in a loop ranging between several > seconds to some minutes. > > Anyone any ideas? I've been staring at this for several days now but > can't find out whats wrong. Can you trigger this issue on linus tree or 5.2-tmp of block tree? Thanks, Ming Lei