On Thu, Sep 10, 2015 at 6:48 AM, Ming Lin <mlin@xxxxxxxxxx> wrote: > These 2 patches added virtio-nvme to kernel and qemu, > basically modified from virtio-blk and nvme code. > > As title said, request for your comments. > > Play it in Qemu with: > -drive file=disk.img,format=raw,if=none,id=D22 \ > -device virtio-nvme-pci,drive=D22,serial=1234,num_queues=4 > > The goal is to have a full NVMe stack from VM guest(virtio-nvme) > to host(vhost_nvme) to LIO NVMe-over-fabrics target. Why is a virtio-nvme guest device needed? I guess there must either be NVMe-only features that you want to pass through, or you think the performance will be significantly better than virtio-blk/virtio-scsi? At first glance it seems like the virtio_nvme guest driver is just another block driver like virtio_blk, so I'm not clear why a virtio-nvme device makes sense. > Now there are lots of duplicated code with linux/nvme-core.c and qemu/nvme.c. > The ideal result is to have a multi level NVMe stack(similar as SCSI). > So we can re-use the nvme code, for example > > .-------------------------. > | NVMe device register | > Upper level | NVMe protocol process | > | | > '-------------------------' > > > > .-----------. .-----------. .------------------. > Lower level | PCIe | | VIRTIO | |NVMe over Fabrics | > | | | | |initiator | > '-----------' '-----------' '------------------' You mentioned LIO and SCSI. How will NVMe over Fabrics be integrated into LIO? If it is mapped to SCSI then using virtio_scsi in the guest and tcm_vhost should work. Please also post virtio draft specifications documenting the virtio device. Stefan _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization