On Nov 22 13:30, Jonathan Derrick wrote: > > > On 11/22/2022 1:26 AM, Klaus Jensen wrote: > > On Nov 21 16:04, Keith Busch wrote: > >> [cc'ing Klaus] > >> > >> On Mon, Nov 21, 2022 at 03:49:45PM -0700, Jonathan Derrick wrote: > >>> On 11/21/2022 3:34 PM, Jonathan Derrick wrote: > >>>> On 11/21/2022 1:55 PM, Keith Busch wrote: > >>>>> On Thu, Nov 17, 2022 at 02:22:10PM -0700, Jonathan Derrick wrote: > >>>>>> I seem to have isolated the error mechanism for older kernels, but 6.2.0-rc2 > >>>>>> reliably segfaults my QEMU instance (something else to look into) and I don't > >>>>>> have any 'real' hardware to test this on at the moment. It looks like several > >>>>>> passthru commands are able to enqueue prior/during/after resetting/connecting. > >>>>> > >>>>> I'm not seeing any problem with the latest nvme-qemu after several dozen > >>>>> iterations of this test case. In that environment, the formats and > >>>>> resets complete practically synchronously with the call, so everything > >>>>> proceeds quickly. Is there anything special I need to change? > >>>>> > >>>> I can still repro this with nvme-fixes tag, so I'll have to dig into it myself > >>> Here's a backtrace: > >>> > >>> Thread 1 "qemu-system-x86" received signal SIGSEGV, Segmentation fault. > >>> [Switching to Thread 0x7ffff7554400 (LWP 531154)] > >>> 0x000055555597a9d5 in nvme_ctrl (req=0x7fffec892780) at ../hw/nvme/nvme.h:539 > >>> 540 return sq->ctrl; > >>> (gdb) backtrace > >>> #0 0x000055555597a9d5 in nvme_ctrl (req=0x7fffec892780) at ../hw/nvme/nvme.h:539 > >>> #1 0x0000555555994360 in nvme_format_bh (opaque=0x5555579dd000) at ../hw/nvme/ctrl.c:5852 > >> > >> Thanks, looks like a race between the admin queue format's bottom half, > >> and the controller reset tearing down that queue. I'll work with Klaus > >> on that qemu side (looks like a well placed qemu_bh_cancel() should do > >> it). > >> > > > > Yuck. Bug located and quelched I think. > > > > Jonathan, please try > > > > https://lore.kernel.org/qemu-devel/20221122081348.49963-2-its@xxxxxxxxxxxxx/ > > > > This fixes the qemu crash, but I still see a "nvme still not live after > > 42 seconds!" resulting from the test. I'm seeing A LOT of invalid > > submission queue doorbell writes: > > > > pci_nvme_ub_db_wr_invalid_sq in nvme_process_db: submission queue doorbell write for nonexistent queue, sqid=0, ignoring > > > > Tested on a 6.1-rc4. > > Good change, just defers it a bit for me: > > Thread 1 "qemu-system-x86" received signal SIGSEGV, Segmentation fault. > [Switching to Thread 0x7ffff7554400 (LWP 559269)] > 0x000055555598922e in nvme_enqueue_req_completion (cq=0x0, req=0x7fffec141310) at ../hw/nvme/ctrl.c:1390 > 1390 assert(cq->cqid == req->sq->cqid); > (gdb) backtrace > #0 0x000055555598922e in nvme_enqueue_req_completion (cq=0x0, req=0x7fffec141310) at ../hw/nvme/ctrl.c:1390 > #1 0x000055555598a7a7 in nvme_misc_cb (opaque=0x7fffec141310, ret=0) at ../hw/nvme/ctrl.c:2002 > #2 0x000055555599448a in nvme_do_format (iocb=0x55555770ccd0) at ../hw/nvme/ctrl.c:5891 > #3 0x00005555559942a9 in nvme_format_ns_cb (opaque=0x55555770ccd0, ret=0) at ../hw/nvme/ctrl.c:5828 > #4 0x0000555555dda018 in blk_aio_complete (acb=0x7fffec1fccd0) at ../block/block-backend.c:1501 > #5 0x0000555555dda2fc in blk_aio_write_entry (opaque=0x7fffec1fccd0) at ../block/block-backend.c:1568 > #6 0x0000555555f506b9 in coroutine_trampoline (i0=-331119632, i1=32767) at ../util/coroutine-ucontext.c:177 > #7 0x00007ffff77c84e0 in __start_context () at ../sysdeps/unix/sysv/linux/x86_64/__start_context.S:91 > #8 0x00007ffff4ff2bd0 in () > #9 0x0000000000000000 in () > Bummer. I'll keep digging.
Attachment:
signature.asc
Description: PGP signature