On Thu, Sep 10, 2015 at 10:02 AM, Ming Lin <mlin@xxxxxxxxxx> wrote: > On Thu, 2015-09-10 at 14:02 +0000, Keith Busch wrote: >> On Wed, 9 Sep 2015, Ming Lin wrote: >> > The goal is to have a full NVMe stack from VM guest(virtio-nvme) >> > to host(vhost_nvme) to LIO NVMe-over-fabrics target. >> > >> > Now there are lots of duplicated code with linux/nvme-core.c and qemu/nvme.c. >> > The ideal result is to have a multi level NVMe stack(similar as SCSI). >> > So we can re-use the nvme code, for example >> > >> > .-------------------------. >> > | NVMe device register | >> > Upper level | NVMe protocol process | >> > | | >> > '-------------------------' >> > >> > >> > >> > .-----------. .-----------. .------------------. >> > Lower level | PCIe | | VIRTIO | |NVMe over Fabrics | >> > | | | | |initiator | >> > '-----------' '-----------' '------------------' >> > >> > todo: >> > - tune performance. Should be as good as virtio-blk/virtio-scsi >> > - support discard/flush/integrity >> > - need Redhat's help for the VIRTIO_ID_NVME pci id >> > - multi level NVMe stack >> >> Hi Ming, > > Hi Keith, > >> >> I'll be out for travel for the next week, so I won't have much time to >> do a proper review till the following week. >> >> I think it'd be better to get this hierarchy setup to make the most reuse >> possible than to have this much code duplication between the existing >> driver and emulated qemu nvme. For better or worse, I think the generic >> nvme layer is where things are going. Are you signed up with the fabrics >> contributors? > > No. How to sign up? Got it. Just saw your previous mail: A message on NVMe Over Fabrics _______________________________________________ Virtualization mailing list Virtualization@xxxxxxxxxxxxxxxxxxxxxxxxxx https://lists.linuxfoundation.org/mailman/listinfo/virtualization