NVMe devices are identified by the vendor specific bits: Bit 3 in OACS (device-wide). Currently made per device, as the nvme namespace is missing in the completion path. Bit 1 in DSM (per-namespace). The OACS change can be removed when the namespace is resolvable from the completion path. >From there, the NVMe specification is extended with the following commands: LightNVM Identify LightNVM Channel identify LightNVM Synchronious/Asynchronious erase LightNVM Get Features LightNVM Set Responsibility LightNVM Get Logical to Physical map LightNVM Get Physical to Logical map The NVMe integration can be tested using Keith Busch NVMe qemu simulator Lwith ightNVM patches on top. This can be found at: https://github/LightNVM/qemu-nvme Contributions in this patch from: Jesper Madsen <jmad@xxxxxx> Signed-off-by: Matias Bjørling <m@xxxxxxxxxxx> --- drivers/block/nvme-core.c | 256 ++++++++++++++++++++++++++++++++++++++++++++-- include/linux/nvme.h | 4 + include/uapi/linux/nvme.h | 57 +++++++++++ 3 files changed, 311 insertions(+), 6 deletions(-) diff --git a/drivers/block/nvme-core.c b/drivers/block/nvme-core.c index 337878b..22319a5 100644 --- a/drivers/block/nvme-core.c +++ b/drivers/block/nvme-core.c @@ -38,6 +38,7 @@ #include <linux/sched.h> #include <linux/slab.h> #include <linux/types.h> +#include <linux/lightnvm.h> #include <scsi/sg.h> #include <asm-generic/io-64-nonatomic-lo-hi.h> @@ -60,6 +61,10 @@ static unsigned char retry_time = 30; module_param(retry_time, byte, 0644); MODULE_PARM_DESC(retry_time, "time in seconds to retry failed I/O"); +static unsigned char force_lightnvm; +module_param(force_lightnvm, byte, 0644); +MODULE_PARM_DESC(force_lightnvm, "force initialization of lightnvm"); + static int nvme_major; module_param(nvme_major, int, 0); @@ -139,8 +144,19 @@ struct nvme_cmd_info { void *ctx; int aborted; struct nvme_queue *nvmeq; + struct nvme_ns *ns; }; +static void host_lba_set(struct nvme_command *cmd, u32 val) +{ + __le32 *cdw12 = &cmd->common.cdw10[2]; + __le32 *cdw13 = &cmd->common.cdw10[3]; + + val = cpu_to_le32(val); + *cdw12 = ((*cdw12) & 0xff00ffff) | ((val & 0xff) << 16); + *cdw13 = ((*cdw13) & 0xff) | (val & 0xffffff00); +} + static int nvme_admin_init_hctx(struct blk_mq_hw_ctx *hctx, void *data, unsigned int hctx_idx) { @@ -405,7 +421,10 @@ static void req_completion(struct nvme_queue *nvmeq, void *ctx, rq_data_dir(req) ? DMA_TO_DEVICE : DMA_FROM_DEVICE); nvme_free_iod(nvmeq->dev, iod); - blk_mq_complete_request(req); + if (nvmeq->dev->oacs & NVME_CTRL_OACS_LIGHTNVM || force_lightnvm) + nvm_complete_request(cmd_rq->ns->nvm_dev, req); + else + blk_mq_complete_request(req); } /* length is in bytes. gfp flags indicates whether we may sleep. */ @@ -576,6 +595,10 @@ static int nvme_queue_rq(struct blk_mq_hw_ctx *hctx, struct request *req) enum dma_data_direction dma_dir; int psegs = req->nr_phys_segments; int result = BLK_MQ_RQ_QUEUE_BUSY; + + if (ns->nvm_dev) + nvm_queue_rq(ns->nvm_dev, req); + /* * Requeued IO has already been prepped */ @@ -591,6 +614,7 @@ static int nvme_queue_rq(struct blk_mq_hw_ctx *hctx, struct request *req) req->special = iod; nvme_set_info(cmd, iod, req_completion); + cmd->ns = ns; if (req->cmd_flags & REQ_DISCARD) { void *range; @@ -895,11 +919,61 @@ static int adapter_delete_sq(struct nvme_dev *dev, u16 sqid) return adapter_delete_queue(dev, nvme_admin_delete_sq, sqid); } +int lnvm_identify(struct nvme_dev *dev, dma_addr_t dma_addr) +{ + struct nvme_command c; + + memset(&c, 0, sizeof(c)); + c.common.opcode = lnvm_admin_identify; + c.common.nsid = cpu_to_le32(0); + c.common.prp1 = cpu_to_le64(dma_addr); + + return nvme_submit_admin_cmd(dev, &c, NULL); +} + +int lnvm_identify_channel(struct nvme_dev *dev, unsigned nsid, + dma_addr_t dma_addr) +{ + struct nvme_command c; + + memset(&c, 0, sizeof(c)); + c.common.opcode = lnvm_admin_identify_channel; + c.common.nsid = cpu_to_le32(nsid); + c.common.cdw10[0] = cpu_to_le32(nsid); + c.common.prp1 = cpu_to_le64(dma_addr); + + return nvme_submit_admin_cmd(dev, &c, NULL); +} + +int lnvm_get_features(struct nvme_dev *dev, unsigned nsid, dma_addr_t dma_addr) +{ + struct nvme_command c; + + memset(&c, 0, sizeof(c)); + c.common.opcode = lnvm_admin_get_features; + c.common.nsid = cpu_to_le32(nsid); + c.common.prp1 = cpu_to_le64(dma_addr); + + return nvme_submit_admin_cmd(dev, &c, NULL); +} + +int lnvm_set_responsibility(struct nvme_dev *dev, unsigned nsid, + dma_addr_t dma_addr) +{ + struct nvme_command c; + + memset(&c, 0, sizeof(c)); + c.common.opcode = lnvm_admin_set_responsibility; + c.common.nsid = cpu_to_le32(nsid); + c.common.prp1 = cpu_to_le64(dma_addr); + + return nvme_submit_admin_cmd(dev, &c, NULL); +} + int nvme_identify(struct nvme_dev *dev, unsigned nsid, unsigned cns, dma_addr_t dma_addr) { struct nvme_command c; - memset(&c, 0, sizeof(c)); c.identify.opcode = nvme_admin_identify; c.identify.nsid = cpu_to_le32(nsid); @@ -1282,6 +1356,90 @@ static int nvme_shutdown_ctrl(struct nvme_dev *dev) return 0; } +static int nvme_nvm_id(struct nvm_dev *nvm_dev, struct nvm_id *nvm_id) +{ + struct nvme_ns *ns = nvm_dev->driver_data; + struct nvme_dev *dev = ns->dev; + struct pci_dev *pdev = dev->pci_dev; + struct nvme_lnvm_id_ctrl *ctrl; + void *mem; + dma_addr_t dma_addr; + unsigned int ret = 0; + + mem = dma_alloc_coherent(&pdev->dev, 4096, &dma_addr, GFP_KERNEL); + if (!mem) + return -ENOMEM; + + ret = lnvm_identify(dev, dma_addr); + if (ret) { + ret = -EIO; + goto out; + } + + ctrl = mem; + nvm_id->ver_id = le16_to_cpu(ctrl->ver_id); + nvm_id->nvm_type = ctrl->nvm_type; + nvm_id->nchannels = le16_to_cpu(ctrl->nchannels); + out: + dma_free_coherent(&pdev->dev, 4096, mem, dma_addr); + return ret; +} + + +static int nvme_nvm_id_chnl(struct nvm_dev *nvm_dev, int chnl_id, + struct nvm_id_chnl *ic) +{ + struct nvme_ns *ns = nvm_dev->driver_data; + struct nvme_dev *dev = ns->dev; + struct pci_dev *pdev = dev->pci_dev; + struct nvme_lnvm_id_chnl *chnl; + void *mem; + dma_addr_t dma_addr; + unsigned int ret = 0; + + mem = dma_alloc_coherent(&pdev->dev, 4096, &dma_addr, GFP_KERNEL); + if (!mem) + return -ENOMEM; + + ret = lnvm_identify_channel(dev, chnl_id, dma_addr); + if (ret) { + ret = -EIO; + goto out; + } + + chnl = mem; + ic->queue_size = le64_to_cpu(chnl->queue_size); + ic->gran_read = le64_to_cpu(chnl->gran_read); + ic->gran_write = le64_to_cpu(chnl->gran_write); + ic->gran_erase = le64_to_cpu(chnl->gran_erase); + ic->oob_size = le64_to_cpu(chnl->oob_size); + ic->t_r = le32_to_cpu(chnl->t_r); + ic->t_sqr = le32_to_cpu(chnl->t_sqr); + ic->t_w = le32_to_cpu(chnl->t_w); + ic->t_sqw = le32_to_cpu(chnl->t_sqw); + ic->t_e = le32_to_cpu(chnl->t_e); + ic->io_sched = chnl->io_sched; + ic->laddr_begin = le64_to_cpu(chnl->laddr_begin); + ic->laddr_end = le64_to_cpu(chnl->laddr_end); + out: + dma_free_coherent(&pdev->dev, 4096, mem, dma_addr); + return ret; +} + +static int nvme_nvm_get_features(struct nvm_dev *dev, + struct nvm_get_features *gf) +{ + gf->rsp[0] = (1 << NVM_RSP_L2P); + gf->rsp[0] |= (1 << NVM_RSP_P2L); + gf->rsp[0] |= (1 << NVM_RSP_GC); + return 0; +} + +static int nvme_nvm_set_rsp(struct nvm_dev *dev, u8 rsp, u8 val) +{ + return NVM_RID_NOT_CHANGEABLE | NVM_DNR; +} + static struct blk_mq_ops nvme_mq_admin_ops = { .queue_rq = nvme_admin_queue_rq, .map_queue = blk_mq_map_queue, @@ -1290,6 +1448,13 @@ static struct blk_mq_ops nvme_mq_admin_ops = { .timeout = nvme_timeout, }; +static struct lightnvm_dev_ops nvme_nvm_dev_ops = { + .identify = nvme_nvm_id, + .identify_channel = nvme_nvm_id_chnl, + .get_features = nvme_nvm_get_features, + .set_responsibility = nvme_nvm_set_rsp, +}; + static struct blk_mq_ops nvme_mq_ops = { .queue_rq = nvme_queue_rq, .map_queue = blk_mq_map_queue, @@ -1455,6 +1620,26 @@ void nvme_unmap_user_pages(struct nvme_dev *dev, int write, put_page(sg_page(&iod->sg[i])); } +static int lnvme_submit_io(struct nvme_ns *ns, struct nvme_user_io *io) +{ + struct nvme_command c; + struct nvme_dev *dev = ns->dev; + + memset(&c, 0, sizeof(c)); + c.rw.opcode = io->opcode; + c.rw.flags = io->flags; + c.rw.nsid = cpu_to_le32(ns->ns_id); + c.rw.slba = cpu_to_le64(io->slba); + c.rw.length = cpu_to_le16(io->nblocks); + c.rw.control = cpu_to_le16(io->control); + c.rw.dsmgmt = cpu_to_le32(io->dsmgmt); + c.rw.reftag = cpu_to_le32(io->reftag); + c.rw.apptag = cpu_to_le16(io->apptag); + c.rw.appmask = cpu_to_le16(io->appmask); + + return nvme_submit_io_cmd(dev, ns, &c, NULL); +} + static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio) { struct nvme_dev *dev = ns->dev; @@ -1481,6 +1666,8 @@ static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio) iod = nvme_map_user_pages(dev, io.opcode & 1, io.addr, length); break; default: + if (ns->nvm_dev) + return lnvme_submit_io(ns, &io); return -EINVAL; } @@ -1498,6 +1685,7 @@ static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio) c.rw.reftag = cpu_to_le32(io.reftag); c.rw.apptag = cpu_to_le16(io.apptag); c.rw.appmask = cpu_to_le16(io.appmask); + host_lba_set(&c, io.host_lba); if (meta_len) { meta_iod = nvme_map_user_pages(dev, io.opcode & 1, io.metadata, @@ -1632,6 +1820,13 @@ static int nvme_ioctl(struct block_device *bdev, fmode_t mode, unsigned int cmd, unsigned long arg) { struct nvme_ns *ns = bdev->bd_disk->private_data; + int ret; + + if (ns->nvm_dev) { + ret = nvm_ioctl(ns->nvm_dev, mode, cmd, arg); + if (ret != -ENOTTY) + return ret; + } switch (cmd) { case NVME_IOCTL_ID: @@ -1655,6 +1850,13 @@ static int nvme_compat_ioctl(struct block_device *bdev, fmode_t mode, unsigned int cmd, unsigned long arg) { struct nvme_ns *ns = bdev->bd_disk->private_data; + int ret; + + if (ns->nvm_dev) { + ret = nvm_ioctl(ns->nvm_dev, mode, cmd, arg); + if (ret != -ENOTTY) + return ret; + } switch (cmd) { case SG_IO: @@ -1756,6 +1958,7 @@ static struct nvme_ns *nvme_alloc_ns(struct nvme_dev *dev, unsigned nsid, struct nvme_id_ns *id, struct nvme_lba_range_type *rt) { struct nvme_ns *ns; + struct nvm_dev *nvm_dev = NULL; struct gendisk *disk; int node = dev_to_node(&dev->pci_dev->dev); int lbaf; @@ -1766,15 +1969,27 @@ static struct nvme_ns *nvme_alloc_ns(struct nvme_dev *dev, unsigned nsid, ns = kzalloc_node(sizeof(*ns), GFP_KERNEL, node); if (!ns) return NULL; + + if (id->nsfeat & NVME_NS_FEAT_LIGHTNVM || force_lightnvm) { + nvm_dev = nvm_alloc(); + if (!nvm_dev) + goto out_free_ns; + + nvm_dev->ops = &nvme_nvm_dev_ops; + + nvm_dev->driver_data = ns; + nvm_dev->drv_cmd_size = dev->tagset.cmd_size - nvm_cmd_size(); + } + ns->queue = blk_mq_init_queue(&dev->tagset); if (!ns->queue) - goto out_free_ns; - queue_flag_set_unlocked(QUEUE_FLAG_DEFAULT, ns->queue); + goto out_free_nvm; queue_flag_set_unlocked(QUEUE_FLAG_NOMERGES, ns->queue); queue_flag_set_unlocked(QUEUE_FLAG_NONROT, ns->queue); queue_flag_set_unlocked(QUEUE_FLAG_SG_GAPS, ns->queue); queue_flag_clear_unlocked(QUEUE_FLAG_IO_STAT, ns->queue); ns->dev = dev; + ns->queue->queuedata = ns; disk = alloc_disk_node(0, node); @@ -1807,8 +2022,24 @@ static struct nvme_ns *nvme_alloc_ns(struct nvme_dev *dev, unsigned nsid, if (dev->oncs & NVME_CTRL_ONCS_DSM) nvme_config_discard(ns); + if (id->nsfeat & NVME_NS_FEAT_LIGHTNVM || force_lightnvm) { + /* Limit to 4K until LightNVM supports multiple IOs */ + blk_queue_max_hw_sectors(ns->queue, 8); + + nvm_dev->q = ns->queue; + nvm_dev->disk = disk; + + if (nvm_init(disk, nvm_dev)) + goto out_put_disk; + + ns->nvm_dev = nvm_dev; + } + return ns; - + out_put_disk: + put_disk(disk); + out_free_nvm: + nvm_free(nvm_dev); out_free_queue: blk_cleanup_queue(ns->queue); out_free_ns: @@ -1954,6 +2185,7 @@ static int nvme_dev_add(struct nvme_dev *dev) ctrl = mem; nn = le32_to_cpup(&ctrl->nn); dev->oncs = le16_to_cpup(&ctrl->oncs); + dev->oacs = le16_to_cpup(&ctrl->oacs); dev->abort_limit = ctrl->acl + 1; dev->vwc = ctrl->vwc; memcpy(dev->serial, ctrl->sn, sizeof(ctrl->sn)); @@ -1983,6 +2215,15 @@ static int nvme_dev_add(struct nvme_dev *dev) dev->tagset.flags = BLK_MQ_F_SHOULD_MERGE; dev->tagset.driver_data = dev; + /* LightNVM is actually per ns, but as the tagset is defined with a set + * of operations for the hole device. It currently is either all or + * no lightnvm compatible name-spaces for a given device. This should + * either be moved toward the nvme_queue_rq function, or allow per ns + * queue_rq function to be specified. + */ + if (dev->oacs & NVME_CTRL_OACS_LIGHTNVM || force_lightnvm) + dev->tagset.cmd_size += nvm_cmd_size(); + if (blk_mq_alloc_tag_set(&dev->tagset)) goto out; @@ -2004,8 +2245,11 @@ static int nvme_dev_add(struct nvme_dev *dev) if (ns) list_add_tail(&ns->list, &dev->namespaces); } - list_for_each_entry(ns, &dev->namespaces, list) + list_for_each_entry(ns, &dev->namespaces, list) { add_disk(ns->disk); + if (ns->nvm_dev) + nvm_add_sysfs(ns->nvm_dev); + } res = 0; out: diff --git a/include/linux/nvme.h b/include/linux/nvme.h index 299e6f5..242ad31 100644 --- a/include/linux/nvme.h +++ b/include/linux/nvme.h @@ -20,6 +20,7 @@ #include <linux/miscdevice.h> #include <linux/kref.h> #include <linux/blk-mq.h> +#include <linux/lightnvm.h> struct nvme_bar { __u64 cap; /* Controller Capabilities */ @@ -100,6 +101,7 @@ struct nvme_dev { u32 max_hw_sectors; u32 stripe_size; u16 oncs; + u16 oacs; u16 abort_limit; u8 vwc; u8 initialized; @@ -120,6 +122,8 @@ struct nvme_ns { int ms; u64 mode_select_num_blocks; u32 mode_select_block_len; + + struct nvm_dev *nvm_dev; }; /* diff --git a/include/uapi/linux/nvme.h b/include/uapi/linux/nvme.h index 29a7d86..965da53 100644 --- a/include/uapi/linux/nvme.h +++ b/include/uapi/linux/nvme.h @@ -85,6 +85,30 @@ struct nvme_id_ctrl { __u8 vs[1024]; }; +struct nvme_lnvm_id_ctrl { + __le16 ver_id; + __u8 nvm_type; + __le16 nchannels; + __u8 unused[4091]; +} __attribute__((packed)); + +struct nvme_lnvm_id_chnl { + __le64 queue_size; + __le64 gran_read; + __le64 gran_write; + __le64 gran_erase; + __le64 oob_size; + __le32 t_r; + __le32 t_sqr; + __le32 t_w; + __le32 t_sqw; + __le32 t_e; + __u8 io_sched; + __le64 laddr_begin; + __le64 laddr_end; + __u8 unused[4034]; +} __attribute__((packed)); + enum { NVME_CTRL_ONCS_COMPARE = 1 << 0, NVME_CTRL_ONCS_WRITE_UNCORRECTABLE = 1 << 1, @@ -123,7 +147,12 @@ struct nvme_id_ns { }; enum { + NVME_CTRL_OACS_LIGHTNVM = 1 << 3, +}; + +enum { NVME_NS_FEAT_THIN = 1 << 0, + NVME_NS_FEAT_LIGHTNVM = 1 << 1, NVME_LBAF_RP_BEST = 0, NVME_LBAF_RP_BETTER = 1, NVME_LBAF_RP_GOOD = 2, @@ -192,6 +221,11 @@ enum nvme_opcode { nvme_cmd_dsm = 0x09, }; +enum lnvme_opcode { + lnvme_cmd_erase_sync = 0x80, + lnvme_cmd_erase_async = 0x81, +}; + struct nvme_common_command { __u8 opcode; __u8 flags; @@ -287,6 +321,15 @@ enum nvme_admin_opcode { nvme_admin_security_recv = 0x82, }; +enum lnvm_admin_opcode { + lnvm_admin_identify = 0xc0, + lnvm_admin_identify_channel = 0xc1, + lnvm_admin_get_features = 0xc2, + lnvm_admin_set_responsibility = 0xc3, + lnvm_admin_get_l2p_tbl = 0xc4, + lnvm_admin_get_p2l_tbl = 0xc5, +}; + enum { NVME_QUEUE_PHYS_CONTIG = (1 << 0), NVME_CQ_IRQ_ENABLED = (1 << 1), @@ -410,6 +453,18 @@ struct nvme_format_cmd { __u32 rsvd11[5]; }; +struct nvme_lnvm_identify { + __u8 opcode; + __u8 flags; + __u16 command_id; + __le32 nsid; + __u64 rsvd[2]; + __le64 prp1; + __le64 prp2; + __le32 cns; + __u32 rsvd11[5]; +}; + struct nvme_command { union { struct nvme_common_command common; @@ -423,6 +478,7 @@ struct nvme_command { struct nvme_format_cmd format; struct nvme_dsm_cmd dsm; struct nvme_abort_cmd abort; + struct nvme_lnvm_identify lnvm_identify; }; }; @@ -487,6 +543,7 @@ struct nvme_user_io { __u32 reftag; __u16 apptag; __u16 appmask; + __u32 host_lba; }; struct nvme_admin_cmd { -- 1.9.1 -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html