[+cc Jon, Dave, Allen, linux-ntb] This patch only touches drivers/pci; the subject line should reflect that to help people know which things to look at. Maybe something like: PCI: endpoint: Support NTB transfer between RC and EP On Mon, Feb 14, 2022 at 11:38:43PM -0600, Frank Li wrote: > Add NTB function driver and virtual PCI BUS and Virtual NTB driver > to implement communication between PCIe RC and PCIe EP devices s/PCI BUS/PCI Bus/ s/PCIe RC/PCIe Root Port/ (I think; the RC itself doesn't have a software representation) > ┌────────────┐ ┌─────────────────────────────────────┐ > │ │ │ │ > ├────────────┤ │ ┌──────────────┤ > │ NTB │ │ │ NTB │ > │ NetDev │ │ │ NetDev │ > ├────────────┤ │ ├──────────────┤ > │ NTB │ │ │ NTB │ > │ Transfer │ │ │ Transfer │ > ├────────────┤ │ ├──────────────┤ > │ │ │ │ │ > │ PCI NTB │ │ │ │ > │ EPF │ │ │ │ > │ Driver │ │ │ PCI Virtual │ > │ │ ├───────────────┐ │ NTB Driver │ > │ │ │ PCI EP NTB │◄────►│ │ > │ │ │ FN Driver │ │ │ > ├────────────┤ ├───────────────┤ ├──────────────┤ > │ │ │ │ │ │ > │ PCI BUS │ ◄─────► │ PCI EP BUS │ │ Virtual PCI │ > │ │ PCI │ │ │ BUS │ > └────────────┘ └───────────────┴──────┴──────────────┘ s/PCI BUS/PCI Bus/ > PCI RC PCI EP s/PCI RC/PCIe Root Port/ ? > This driver include 3 part: > 1 PCI EP NTB function driver > 2 Virtual PCI bus > 3 PCI virutal NTB driver, which is loaded only by above virtual pci bus s/include 3 part/includes 3 parts/ s/virutal/virtual/ s/pci bus/PCI bus/ > Signed-off-by: Frank Li <Frank.Li@xxxxxxx> > --- > drivers/pci/endpoint/functions/Kconfig | 11 + > drivers/pci/endpoint/functions/Makefile | 1 + > drivers/pci/endpoint/functions/pci-epf-vntb.c | 1425 +++++++++++++++++ > 3 files changed, 1437 insertions(+) > create mode 100644 drivers/pci/endpoint/functions/pci-epf-vntb.c > > diff --git a/drivers/pci/endpoint/functions/Kconfig b/drivers/pci/endpoint/functions/Kconfig > index 5f1242ca2f4e4..362555b024e8f 100644 > --- a/drivers/pci/endpoint/functions/Kconfig > +++ b/drivers/pci/endpoint/functions/Kconfig > @@ -25,3 +25,14 @@ config PCI_EPF_NTB > device tree. > > If in doubt, say "N" to disable Endpoint NTB driver. > + > +config PCI_EPF_VNTB > + tristate "PCI Endpoint NTB driver" > + depends on PCI_ENDPOINT > + select CONFIGFS_FS > + help > + Select this configuration option to enable the Non-Transparent > + Bridge (NTB) driver for PCI Endpoint. NTB driver implements NTB > + between PCI host and PCIe Endpoint. "PCI Endpoint" vs "PCIe Endpoint". Pick one. It seems like you require an upstream bridge, i.e., a Root Port, so maybe "PCIe Root Port" and "PCIe Endpoint" is what you want? > + If in doubt, say "N" to disable Endpoint NTB driver. > diff --git a/drivers/pci/endpoint/functions/Makefile b/drivers/pci/endpoint/functions/Makefile > index 96ab932a537a2..5c13001deaba1 100644 > --- a/drivers/pci/endpoint/functions/Makefile > +++ b/drivers/pci/endpoint/functions/Makefile > @@ -5,3 +5,4 @@ > > obj-$(CONFIG_PCI_EPF_TEST) += pci-epf-test.o > obj-$(CONFIG_PCI_EPF_NTB) += pci-epf-ntb.o > +obj-$(CONFIG_PCI_EPF_VNTB) += pci-epf-vntb.o > diff --git a/drivers/pci/endpoint/functions/pci-epf-vntb.c b/drivers/pci/endpoint/functions/pci-epf-vntb.c > new file mode 100644 > index 0000000000000..ebf7e243eefa4 > --- /dev/null > +++ b/drivers/pci/endpoint/functions/pci-epf-vntb.c > @@ -0,0 +1,1425 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Endpoint Function Driver to implement Non-Transparent Bridge functionality > + * Between PCI RC and EP > + * > + * Copyright (C) 2020 Texas Instruments > + * Copyright (C) 2022 NXP > + * > + * Based on pci-epf-ntb.c > + * Author: Frank Li <Frank.Li@xxxxxxx> > + * Author: Kishon Vijay Abraham I <kishon@xxxxxx> > + */ > + > +/** > + * +------------+ +---------------------------------------+ > + * | | | | > + * +------------+ | +--------------+ > + * | NTB | | | NTB | > + * | NetDev | | | NetDev | > + * +------------+ | +--------------+ > + * | NTB | | | NTB | > + * | Transfer | | | Transfer | > + * +------------+ | +--------------+ > + * | | | | | > + * | PCI NTB | | | | > + * | EPF | | | | > + * | Driver | | | PCI Virtual | > + * | | +---------------+ | NTB Driver | > + * | | | PCI EP NTB |<------>| | > + * | | | FN Driver | | | > + * +------------+ +---------------+ +--------------+ > + * | | | | | | > + * | PCI BUS | <-----> | PCI EP BUS | | Virtual PCI | > + * | | PCI | | | BUS | > + * +------------+ +---------------+--------+--------------+ > + * PCI RC PCI EP s/PCI RC/PCIe Root Port/ ? > + */ > + > +#include <linux/delay.h> > +#include <linux/io.h> > +#include <linux/module.h> > +#include <linux/slab.h> > + > +#include <linux/pci-epc.h> > +#include <linux/pci-epf.h> > +#include <linux/ntb.h> > + > +static struct workqueue_struct *kpcintb_workqueue; > + > +#define COMMAND_CONFIGURE_DOORBELL 1 > +#define COMMAND_TEARDOWN_DOORBELL 2 > +#define COMMAND_CONFIGURE_MW 3 > +#define COMMAND_TEARDOWN_MW 4 > +#define COMMAND_LINK_UP 5 > +#define COMMAND_LINK_DOWN 6 > + > +#define COMMAND_STATUS_OK 1 > +#define COMMAND_STATUS_ERROR 2 > + > +#define LINK_STATUS_UP BIT(0) > + > +#define SPAD_COUNT 64 > +#define DB_COUNT 4 > +#define NTB_MW_OFFSET 2 > +#define DB_COUNT_MASK GENMASK(15, 0) > +#define MSIX_ENABLE BIT(16) > +#define MAX_DB_COUNT 32 > +#define MAX_MW 4 > + > +#define VNTB_VID 0x1957 This looks like PCI_VENDOR_ID_NXP (or PCI_VENDOR_ID_FREESCALE). How did you choose that, why is it safe to use, and why don't use use PCI_VENDOR_ID_NXP instead? > +#define VNTB_PID 0x080A > + > +enum epf_ntb_bar { > + BAR_CONFIG, > + BAR_DB, > + BAR_MW0, > + BAR_MW1, > + BAR_MW2, > +}; > + > +/* > + * +--------------------------------------------------+ Base > + * | | > + * | | > + * | | > + * | Common Control Register | > + * | | > + * | | > + * | | > + * +-----------------------+--------------------------+ Base+span_offset > + * | | | > + * | Peer Span Space | Span Space | > + * | | | > + * | | | > + * +-----------------------+--------------------------+ Base+span_offset > + * | | | +span_count * 4 > + * | | | > + * | Span Space | Peer Span Space | > + * | | | > + * +-----------------------+--------------------------+ > + * Virtual PCI Pcie Endpoint > + * NTB Driver NTB Driver s/Pcie/PCIe/ > + */ > +struct epf_ntb_ctrl { > + u32 command; > + u32 argument; > + u16 command_status; > + u16 link_status; > + u32 topology; > + u64 addr; > + u64 size; > + u32 num_mws; > + u32 reserved; > + u32 spad_offset; > + u32 spad_count; > + u32 db_entry_size; > + u32 db_data[MAX_DB_COUNT]; > + u32 db_offset[MAX_DB_COUNT]; > +} __packed; > + > +struct epf_ntb { > + struct ntb_dev ntb; > + struct pci_epf *epf; > + struct config_group group; > + > + u32 num_mws; > + u32 db_count; > + u32 spad_count; > + u64 mws_size[MAX_MW]; > + u64 db; > + u32 vbus_number; > + > + bool linkup; > + u32 spad_size; > + > + enum pci_barno epf_ntb_bar[6]; > + > + struct epf_ntb_ctrl *reg; > + > + phys_addr_t epf_db_phy; > + void __iomem *epf_db; > + > + phys_addr_t vpci_mw_phy[MAX_MW]; > + void __iomem *vpci_mw_addr[MAX_MW]; > + > + struct delayed_work cmd_handler; > +}; > + > +#define to_epf_ntb(epf_group) container_of((epf_group), struct epf_ntb, group) > +#define ntb_ndev(__ntb) container_of(__ntb, struct epf_ntb, ntb) > + > +static struct pci_epf_header epf_ntb_header = { > + .vendorid = PCI_ANY_ID, > + .deviceid = PCI_ANY_ID, > + .baseclass_code = PCI_BASE_CLASS_MEMORY, > + .interrupt_pin = PCI_INTERRUPT_INTA, > +}; > + > +/** > + * epf_ntb_link_up() - Raise link_up interrupt to Virtual Host > + * @ntb: NTB device that facilitates communication between HOST and VHOST > + * @link_up: true or false indicating Link is UP or Down > + * > + * Once NTB function in HOST invoke ntb_link_enable(), > + * this NTB function driver will trigger a link event to vhost. > + */ > +static int epf_ntb_link_up(struct epf_ntb *ntb, bool link_up) > +{ > + if (link_up) > + ntb->reg->link_status |= LINK_STATUS_UP; > + else > + ntb->reg->link_status &= ~LINK_STATUS_UP; > + > + ntb_link_event(&ntb->ntb); > + return 0; > +} > + > +/** > + * epf_ntb_configure_mw() - Configure the Outbound Address Space for vhost > + * to access the memory window of host > + * @ntb: NTB device that facilitates communication between host and vhost > + * @mw: Index of the memory window (either 0, 1, 2 or 3) > + * > + * EP Outbound Window > + * +--------+ +-----------+ > + * | | | | > + * | | | | > + * | | | | > + * | | | | > + * | | +-----------+ > + * | Virtual| | Memory Win| > + * | NTB | -----------> | | > + * | Driver | | | > + * | | +-----------+ > + * | | | | > + * | | | | > + * +--------+ +-----------+ > + * VHost PCI EP > + */ > +static int epf_ntb_configure_mw(struct epf_ntb *ntb, u32 mw) > +{ > + phys_addr_t phys_addr; > + u8 func_no, vfunc_no; > + u64 addr, size; > + int ret = 0; > + > + phys_addr = ntb->vpci_mw_phy[mw]; > + addr = ntb->reg->addr; > + size = ntb->reg->size; > + > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + ret = pci_epc_map_addr(ntb->epf->epc, func_no, vfunc_no, phys_addr, addr, size); > + if (ret) > + dev_err(&ntb->epf->epc->dev, > + "intf: Failed to map memory window %d address\n", mw); Lots of these messages start with "intf:". What is that telling us? Is it useful? > + return ret; > +} > + > +/** > + * epf_ntb_teardown_mw() - Teardown the configured OB ATU > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * @mw: Index of the memory window (either 0, 1, 2 or 3) > + * > + * Teardown the configured OB ATU configured in epf_ntb_configure_mw() using > + * pci_epc_unmap_addr() > + */ > +static void epf_ntb_teardown_mw(struct epf_ntb *ntb, u32 mw) > +{ > + pci_epc_unmap_addr(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + ntb->vpci_mw_phy[mw]); > +} > + > +/** > + * epf_ntb_cmd_handler() - Handle commands provided by the NTB Host > + * @work: work_struct for the epf_ntb_epc > + * > + * Workqueue function that gets invoked for the two epf_ntb_epc > + * periodically (once every 5ms) to see if it has received any commands > + * from NTB host. The host can send commands to configure doorbell or > + * configure memory window or to update link status. > + */ > +static void epf_ntb_cmd_handler(struct work_struct *work) > +{ > + struct epf_ntb_ctrl *ctrl; > + u32 command, argument; > + struct epf_ntb *ntb; > + struct device *dev; > + int ret; > + int i; > + > + ntb = container_of(work, struct epf_ntb, cmd_handler.work); > + > + for (i = 1; i < ntb->db_count; i++) { > + if (readl(ntb->epf_db + i * 4)) { > + if (readl(ntb->epf_db + i * 4)) > + ntb->db |= 1 << (i - 1); > + > + ntb_db_event(&ntb->ntb, i); > + writel(0, ntb->epf_db + i * 4); > + } > + } > + > + ctrl = ntb->reg; > + command = ctrl->command; > + if (!command) > + goto reset_handler; > + argument = ctrl->argument; > + > + ctrl->command = 0; > + ctrl->argument = 0; > + > + ctrl = ntb->reg; > + dev = &ntb->epf->dev; > + > + switch (command) { > + case COMMAND_CONFIGURE_DOORBELL: > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_TEARDOWN_DOORBELL: > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_CONFIGURE_MW: > + ret = epf_ntb_configure_mw(ntb, argument); > + if (ret < 0) > + ctrl->command_status = COMMAND_STATUS_ERROR; > + else > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_TEARDOWN_MW: > + epf_ntb_teardown_mw(ntb, argument); > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + case COMMAND_LINK_UP: > + ntb->linkup = true; > + ret = epf_ntb_link_up(ntb, true); > + if (ret < 0) > + ctrl->command_status = COMMAND_STATUS_ERROR; > + else > + ctrl->command_status = COMMAND_STATUS_OK; > + goto reset_handler; > + case COMMAND_LINK_DOWN: > + ntb->linkup = false; > + ret = epf_ntb_link_up(ntb, false); > + if (ret < 0) > + ctrl->command_status = COMMAND_STATUS_ERROR; > + else > + ctrl->command_status = COMMAND_STATUS_OK; > + break; > + default: > + dev_err(dev, "intf UNKNOWN command: %d\n", command); > + break; > + } > + > +reset_handler: > + queue_delayed_work(kpcintb_workqueue, &ntb->cmd_handler, > + msecs_to_jiffies(5)); > +} > + > +/** > + * epf_ntb_config_sspad_bar_clear() - Clear Config + Self scratchpad BAR > + * @ntb_epc: EPC associated with one of the HOST which holds peer's outbound > + * address. > + * > + * Clear BAR0 of EP CONTROLLER 1 which contains the HOST1's config and > + * self scratchpad region (removes inbound ATU configuration). While BAR0 is > + * the default self scratchpad BAR, an NTB could have other BARs for self > + * scratchpad (because of reserved BARs). This function can get the exact BAR > + * used for self scratchpad from epf_ntb_bar[BAR_CONFIG]. > + * > + * Please note the self scratchpad region and config region is combined to > + * a single region and mapped using the same BAR. Also note HOST2's peer > + * scratchpad is HOST1's self scratchpad. > + */ > +static void epf_ntb_config_sspad_bar_clear(struct epf_ntb *ntb) > +{ > + struct pci_epf_bar *epf_bar; > + enum pci_barno barno; > + > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + epf_bar = &ntb->epf->bar[barno]; > + > + pci_epc_clear_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); > +} > + > +/** > + * epf_ntb_config_sspad_bar_set() - Set Config + Self scratchpad BAR > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + * Map BAR0 of EP CONTROLLER 1 which contains the HOST1's config and > + * self scratchpad region. > + * > + * Please note the self scratchpad region and config region is combined to > + * a single region and mapped using the same BAR. > + */ > +static int epf_ntb_config_sspad_bar_set(struct epf_ntb *ntb) > +{ > + struct pci_epf_bar *epf_bar; > + enum pci_barno barno; > + u8 func_no, vfunc_no; > + struct device *dev; > + int ret; > + > + dev = &ntb->epf->dev; > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + epf_bar = &ntb->epf->bar[barno]; > + > + ret = pci_epc_set_bar(ntb->epf->epc, func_no, vfunc_no, epf_bar); > + if (ret) { > + dev_err(dev, "inft: Config/Status/SPAD BAR set failed\n"); > + return ret; > + } > + return 0; > +} > + > +/** > + * epf_ntb_config_spad_bar_free() - Free the physical memory associated with > + * config + scratchpad region > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + */ > +static void epf_ntb_config_spad_bar_free(struct epf_ntb *ntb) > +{ > + enum pci_barno barno; > + > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + pci_epf_free_space(ntb->epf, ntb->reg, barno, 0); > +} > + > +/** > + * epf_ntb_config_spad_bar_alloc() - Allocate memory for config + scratchpad > + * region > + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 > + * > + * Allocate the Local Memory mentioned in the above diagram. The size of > + * CONFIG REGION is sizeof(struct epf_ntb_ctrl) and size of SCRATCHPAD REGION > + * is obtained from "spad-count" configfs entry. > + */ > +static int epf_ntb_config_spad_bar_alloc(struct epf_ntb *ntb) > +{ > + size_t align; > + enum pci_barno barno; > + struct epf_ntb_ctrl *ctrl; > + u32 spad_size, ctrl_size; > + u64 size; > + struct pci_epf *epf = ntb->epf; > + struct device *dev = &epf->dev; > + u32 spad_count; > + void *base; > + int i; > + const struct pci_epc_features *epc_features = pci_epc_get_features(epf->epc, > + epf->func_no, > + epf->vfunc_no); > + barno = ntb->epf_ntb_bar[BAR_CONFIG]; > + size = epc_features->bar_fixed_size[barno]; > + align = epc_features->align; > + > + if ((!IS_ALIGNED(size, align))) > + return -EINVAL; > + > + spad_count = ntb->spad_count; > + > + ctrl_size = sizeof(struct epf_ntb_ctrl); > + spad_size = 2 * spad_count * 4; > + > + if (!align) { > + ctrl_size = roundup_pow_of_two(ctrl_size); > + spad_size = roundup_pow_of_two(spad_size); > + } else { > + ctrl_size = ALIGN(ctrl_size, align); > + spad_size = ALIGN(spad_size, align); > + } > + > + if (!size) > + size = ctrl_size + spad_size; > + else if (size < ctrl_size + spad_size) > + return -EINVAL; > + > + base = pci_epf_alloc_space(epf, size, barno, align, 0); > + if (!base) { > + dev_err(dev, "intf: Config/Status/SPAD alloc region fail\n"); > + return -ENOMEM; > + } > + > + ntb->reg = base; > + > + ctrl = ntb->reg; > + ctrl->spad_offset = ctrl_size; > + > + ctrl->spad_count = spad_count; > + ctrl->num_mws = ntb->num_mws; > + ntb->spad_size = spad_size; > + > + ctrl->db_entry_size = 4; > + > + for (i = 0; i < ntb->db_count; i++) { > + ntb->reg->db_data[i] = 1 + i; > + ntb->reg->db_offset[i] = 0; > + } > + > + return 0; > +} > + > +/** > + * epf_ntb_configure_interrupt() - Configure MSI/MSI-X capaiblity > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + * Configure MSI/MSI-X capability for each interface with number of > + * interrupts equal to "db_count" configfs entry. > + */ > +static int epf_ntb_configure_interrupt(struct epf_ntb *ntb) > +{ > + const struct pci_epc_features *epc_features; > + bool msix_capable, msi_capable; > + u8 func_no, vfunc_no; > + struct device *dev; > + u32 db_count; > + int ret; > + > + dev = &ntb->epf->dev; > + > + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); > + msix_capable = epc_features->msix_capable; > + msi_capable = epc_features->msi_capable; > + > + if (!(msix_capable || msi_capable)) { I don't think the "msix_capable" and "msi_capable" local variables really add any readability since they're only used once. Not sure about "func_no", "vfunc_no", either. If you keep them, you could at least assign them earlier and use them in the pci_epc_get_features() call. > + dev_err(dev, "MSI or MSI-X is required for doorbell\n"); > + return -EINVAL; > + } > + > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + db_count = ntb->db_count; > + if (db_count > MAX_DB_COUNT) { > + dev_err(dev, "DB count cannot be more than %d\n", MAX_DB_COUNT); > + return -EINVAL; > + } > + > + ntb->db_count = db_count; > + > + if (msi_capable) { > + ret = pci_epc_set_msi(ntb->epf->epc, func_no, vfunc_no, 16); > + if (ret) { > + dev_err(dev, "intf: MSI configuration failed\n"); > + return ret; > + } > + } > + > + return 0; > +} > + > +/** > + * epf_ntb_db_bar_init() - Configure Doorbell window BARs > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * Spurious blank line. > + */ > +static int epf_ntb_db_bar_init(struct epf_ntb *ntb) > +{ > + const struct pci_epc_features *epc_features; > + u32 align; > + struct device *dev = &ntb->epf->dev; > + int ret; > + struct pci_epf_bar *epf_bar; > + void __iomem *mw_addr; > + enum pci_barno barno; > + size_t size = 4 * ntb->db_count; > + > + epc_features = pci_epc_get_features(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no); > + align = epc_features->align; > + > + if (size < 128) > + size = 128; > + > + if (align) > + size = ALIGN(size, align); > + else > + size = roundup_pow_of_two(size); > + > + barno = ntb->epf_ntb_bar[BAR_DB]; > + > + mw_addr = pci_epf_alloc_space(ntb->epf, size, barno, align, 0); > + if (!mw_addr) { > + dev_err(dev, "intf: Failed to allocate OB address\n"); > + return -ENOMEM; > + } > + > + ntb->epf_db = mw_addr; > + > + epf_bar = &ntb->epf->bar[barno]; > + > + ret = pci_epc_set_bar(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no, epf_bar); > + if (ret) { > + dev_err(dev, "intf: DoorBell BAR set failed\n"); s/DoorBell/Doorbell/ > + goto err_alloc_peer_mem; > + } > + return ret; > + > +err_alloc_peer_mem: > + pci_epc_mem_free_addr(ntb->epf->epc, epf_bar->phys_addr, mw_addr, epf_bar->size); > + return -1; > +} > + > +/** > + * epf_ntb_db_bar_clear() - Clear doorbell BAR and free memory > + * allocated in peers outbound address space s/peers/peer's/ > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + */ > +static void epf_ntb_db_bar_clear(struct epf_ntb *ntb) > +{ > + enum pci_barno barno; > + > + barno = ntb->epf_ntb_bar[BAR_DB]; > + pci_epf_free_space(ntb->epf, ntb->epf_db, barno, 0); > + pci_epc_clear_bar(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + &ntb->epf->bar[barno]); > +} > + > +/** > + * epf_ntb_mw_bar_init() - Configure Memory window BARs > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + */ > +static int epf_ntb_mw_bar_init(struct epf_ntb *ntb) > +{ > + int ret = 0; > + int i; > + u64 size; > + enum pci_barno barno; > + struct device *dev = &ntb->epf->dev; > + > + for (i = 0; i < ntb->num_mws; i++) { > + Spurious blank line. > + size = ntb->mws_size[i]; > + > + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; > + > + ntb->epf->bar[barno].barno = barno; > + ntb->epf->bar[barno].size = size; > + ntb->epf->bar[barno].addr = 0; > + ntb->epf->bar[barno].phys_addr = 0; > + ntb->epf->bar[barno].flags |= upper_32_bits(size) ? > + PCI_BASE_ADDRESS_MEM_TYPE_64 : > + PCI_BASE_ADDRESS_MEM_TYPE_32; > + > + ret = pci_epc_set_bar(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + &ntb->epf->bar[barno]); > + if (ret) { > + dev_err(dev, "intf: MW set failed\n"); > + goto err_alloc_mem; > + } > + > + /* allocate epc outbound memory windows to vpci vntb device */ s/allocate/Allocate/ to match other comments s/epc/EPC/ > + ntb->vpci_mw_addr[i] = pci_epc_mem_alloc_addr(ntb->epf->epc, > + &ntb->vpci_mw_phy[i], > + size); > + if (!ntb->vpci_mw_addr[i]) { > + dev_err(dev, "Failed to allocate source address\n"); > + goto err_alloc_mem; > + } > + } > + > + return ret; > +err_alloc_mem: > + return ret; > +} > + > +/** > + * epf_ntb_mw_bar_clear() - Clear Memory window BARs > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * Spurious blank line. > + */ > +static void epf_ntb_mw_bar_clear(struct epf_ntb *ntb) > +{ > + enum pci_barno barno; > + int i; > + > + for (i = 0; i < ntb->num_mws; i++) { > + barno = ntb->epf_ntb_bar[BAR_MW0 + i]; > + pci_epc_clear_bar(ntb->epf->epc, > + ntb->epf->func_no, > + ntb->epf->vfunc_no, > + &ntb->epf->bar[barno]); > + > + pci_epc_mem_free_addr(ntb->epf->epc, > + ntb->vpci_mw_phy[i], > + ntb->vpci_mw_addr[i], > + ntb->mws_size[i]); > + } > +} > + > +/** > + * epf_ntb_epc_destroy() - Cleanup NTB EPC interface > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * > + * Wrapper for epf_ntb_epc_destroy_interface() to cleanup all the NTB interfaces > + */ > +static void epf_ntb_epc_destroy(struct epf_ntb *ntb) > +{ > + pci_epc_remove_epf(ntb->epf->epc, ntb->epf, 0); > + pci_epc_put(ntb->epf->epc); > +} > + > +/** > + * epf_ntb_init_epc_bar() - Identify BARs to be used for each of the NTB > + * constructs (scratchpad region, doorbell, memorywindow) > + * @ntb: NTB device that facilitates communication between HOST and vHOST > + * Spurious blank line. > + */ > +static int epf_ntb_init_epc_bar(struct epf_ntb *ntb) > +{ > + const struct pci_epc_features *epc_features; > + enum pci_barno barno; > + enum epf_ntb_bar bar; > + struct device *dev; > + u32 num_mws; > + int i; > + > + barno = BAR_0; > + num_mws = ntb->num_mws; > + dev = &ntb->epf->dev; > + epc_features = pci_epc_get_features(ntb->epf->epc, ntb->epf->func_no, ntb->epf->vfunc_no); > + > + /* These are required BARs which are mandatory for NTB functionality */ > + for (bar = BAR_CONFIG; bar <= BAR_MW0; bar++, barno++) { > + barno = pci_epc_get_next_free_bar(epc_features, barno); > + if (barno < 0) { > + dev_err(dev, "intf: Fail to get NTB function BAR\n"); > + return barno; > + } > + ntb->epf_ntb_bar[bar] = barno; > + } > + > + /* These are optional BARs which don't impact NTB functionality */ > + for (bar = BAR_MW1, i = 1; i < num_mws; bar++, barno++, i++) { > + barno = pci_epc_get_next_free_bar(epc_features, barno); > + if (barno < 0) { > + ntb->num_mws = i; > + dev_dbg(dev, "BAR not available for > MW%d\n", i + 1); > + } > + ntb->epf_ntb_bar[bar] = barno; > + } > + > + return 0; > +} > + > +/** > + * epf_ntb_epc_init() - Initialize NTB interface > + * @ntb: NTB device that facilitates communication between HOST and vHOST2 > + * > + * Wrapper to initialize a particular EPC interface and start the workqueue > + * to check for commands from host. This function will write to the > + * EP controller HW for configuring it. > + */ > +static int epf_ntb_epc_init(struct epf_ntb *ntb) > +{ > + u8 func_no, vfunc_no; > + struct pci_epc *epc; > + struct pci_epf *epf; > + struct device *dev; > + int ret; > + > + epf = ntb->epf; > + dev = &epf->dev; > + epc = epf->epc; > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + ret = epf_ntb_config_sspad_bar_set(ntb); > + if (ret) { > + dev_err(dev, "intf: Config/self SPAD BAR init failed"); > + return ret; > + } > + > + ret = epf_ntb_configure_interrupt(ntb); > + if (ret) { > + dev_err(dev, "intf: Interrupt configuration failed\n"); > + goto err_config_interrupt; > + } > + > + ret = epf_ntb_db_bar_init(ntb); > + if (ret) { > + dev_err(dev, "intf: DB BAR init failed\n"); > + goto err_db_bar_init; > + } > + > + ret = epf_ntb_mw_bar_init(ntb); > + if (ret) { > + dev_err(dev, "intf: MW BAR init failed\n"); > + goto err_mw_bar_init; > + } > + > + if (vfunc_no <= 1) { > + ret = pci_epc_write_header(epc, func_no, vfunc_no, epf->header); > + if (ret) { > + dev_err(dev, "intf: Configuration header write failed\n"); > + goto err_write_header; > + } > + } > + > + INIT_DELAYED_WORK(&ntb->cmd_handler, epf_ntb_cmd_handler); > + queue_work(kpcintb_workqueue, &ntb->cmd_handler.work); > + > + return 0; > + > +err_write_header: > + epf_ntb_mw_bar_clear(ntb); > +err_mw_bar_init: > + epf_ntb_db_bar_clear(ntb); > +err_db_bar_init: > +err_config_interrupt: > + epf_ntb_config_sspad_bar_clear(ntb); > + > + return ret; > +} > + > + > +/** > + * epf_ntb_epc_cleanup() - Cleanup all NTB interfaces > + * @ntb: NTB device that facilitates communication between HOST1 and HOST2 > + * > + * Wrapper to cleanup all NTB interfaces. > + */ > +static void epf_ntb_epc_cleanup(struct epf_ntb *ntb) > +{ > + epf_ntb_db_bar_clear(ntb); > + epf_ntb_mw_bar_clear(ntb); > +} > + > +#define EPF_NTB_R(_name) \ > +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ > + char *page) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + \ > + return sprintf(page, "%d\n", ntb->_name); \ > +} > + > +#define EPF_NTB_W(_name) \ > +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ > + const char *page, size_t len) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + u32 val; \ > + int ret; \ > + \ > + ret = kstrtou32(page, 0, &val); \ > + if (ret) \ > + return ret; \ > + \ > + ntb->_name = val; \ > + \ > + return len; \ > +} > + > +#define EPF_NTB_MW_R(_name) \ > +static ssize_t epf_ntb_##_name##_show(struct config_item *item, \ > + char *page) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + int win_no; \ > + \ > + sscanf(#_name, "mw%d", &win_no); \ > + \ > + return sprintf(page, "%lld\n", ntb->mws_size[win_no - 1]); \ > +} > + > +#define EPF_NTB_MW_W(_name) \ > +static ssize_t epf_ntb_##_name##_store(struct config_item *item, \ > + const char *page, size_t len) \ > +{ \ > + struct config_group *group = to_config_group(item); \ > + struct epf_ntb *ntb = to_epf_ntb(group); \ > + struct device *dev = &ntb->epf->dev; \ > + int win_no; \ > + u64 val; \ > + int ret; \ > + \ > + ret = kstrtou64(page, 0, &val); \ > + if (ret) \ > + return ret; \ > + \ > + if (sscanf(#_name, "mw%d", &win_no) != 1) \ > + return -EINVAL; \ > + \ > + if (ntb->num_mws < win_no) { \ > + dev_err(dev, "Invalid num_nws: %d value\n", ntb->num_mws); \ > + return -EINVAL; \ > + } \ > + \ > + ntb->mws_size[win_no - 1] = val; \ > + \ > + return len; \ > +} > + > +static ssize_t epf_ntb_num_mws_store(struct config_item *item, > + const char *page, size_t len) > +{ > + struct config_group *group = to_config_group(item); > + struct epf_ntb *ntb = to_epf_ntb(group); > + u32 val; > + int ret; > + > + ret = kstrtou32(page, 0, &val); > + if (ret) > + return ret; > + > + if (val > MAX_MW) > + return -EINVAL; > + > + ntb->num_mws = val; > + > + return len; > +} > + > +EPF_NTB_R(spad_count) > +EPF_NTB_W(spad_count) > +EPF_NTB_R(db_count) > +EPF_NTB_W(db_count) > +EPF_NTB_R(num_mws) > +EPF_NTB_R(vbus_number) > +EPF_NTB_W(vbus_number) > +EPF_NTB_MW_R(mw1) > +EPF_NTB_MW_W(mw1) > +EPF_NTB_MW_R(mw2) > +EPF_NTB_MW_W(mw2) > +EPF_NTB_MW_R(mw3) > +EPF_NTB_MW_W(mw3) > +EPF_NTB_MW_R(mw4) > +EPF_NTB_MW_W(mw4) > + > +CONFIGFS_ATTR(epf_ntb_, spad_count); > +CONFIGFS_ATTR(epf_ntb_, db_count); > +CONFIGFS_ATTR(epf_ntb_, num_mws); > +CONFIGFS_ATTR(epf_ntb_, mw1); > +CONFIGFS_ATTR(epf_ntb_, mw2); > +CONFIGFS_ATTR(epf_ntb_, mw3); > +CONFIGFS_ATTR(epf_ntb_, mw4); > +CONFIGFS_ATTR(epf_ntb_, vbus_number); > + > +static struct configfs_attribute *epf_ntb_attrs[] = { > + &epf_ntb_attr_spad_count, > + &epf_ntb_attr_db_count, > + &epf_ntb_attr_num_mws, > + &epf_ntb_attr_mw1, > + &epf_ntb_attr_mw2, > + &epf_ntb_attr_mw3, > + &epf_ntb_attr_mw4, > + &epf_ntb_attr_vbus_number, > + NULL, > +}; > + > +static const struct config_item_type ntb_group_type = { > + .ct_attrs = epf_ntb_attrs, > + .ct_owner = THIS_MODULE, > +}; > + > +/** > + * epf_ntb_add_cfs() - Add configfs directory specific to NTB > + * @epf: NTB endpoint function device > + * @group: A pointer to the config_group structure referencing a group of > + * config_items of a specific type that belong to a specific sub-system. > + * > + * Add configfs directory specific to NTB. This directory will hold > + * NTB specific properties like db_count, spad_count, num_mws etc., > + */ > +static struct config_group *epf_ntb_add_cfs(struct pci_epf *epf, > + struct config_group *group) > +{ > + struct epf_ntb *ntb = epf_get_drvdata(epf); > + struct config_group *ntb_group = &ntb->group; > + struct device *dev = &epf->dev; > + > + config_group_init_type_name(ntb_group, dev_name(dev), &ntb_group_type); > + > + return ntb_group; > +} > + > +/*==== virtual PCI bus driver, which only load virutal ntb pci driver ====*/ s/virutal ntb pci/virtual NTB PCI/ > +#define VPCI_BUS_NUM 0x10 How did you pick this number and how do we know it is available? > +uint32_t pci_space[] = { "static u32"? > + (VNTB_VID | (VNTB_PID << 16)), //DeviceID, Vendor ID > + 0, // status, Command > + 0xffffffff, // Class code, subclass, prog if, revision id > + 0x40, //bist, header type, latency Timer, cache line size > + 0, //bar 0 > + 0, //bar 1 > + 0, //bar 2 > + 0, //bar 3 > + 0, //bar 4 > + 0, //bar 5 s/bar/BAR/ > + 0, //cardbus cis point > + 0, //Subsystem ID Subystem vendor id > + 0, //ROM Base Address > + 0, //Reserved, Cap. Point > + 0, //Reserved, > + 0, //Max Lat, Min Gnt, interrupt pin, interrupt line Use /* */ instead of //, like the rest of drivers/pci/ > +}; > + > +int pci_read(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 *val) > +{ > + if (devfn == 0) { > + memcpy(val, ((uint8_t *)pci_space) + where, size); "u8 *"? > + return 0; > + } > + return -1; These should return PCIBIOS_SUCCESSFUL or PCIBIOS_DEVICE_NOT_FOUND. > +} > + > +int pci_write(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 val) > +{ > + return 0; > +} > + > +struct pci_ops vpci_ops = { > + .read = pci_read, > + .write = pci_write, > +}; > + > +static int vpci_bus(void *sysdata) The function name should say something about what it does. Maybe vpci_scan_bus() or something? > +{ > + struct pci_bus *vpci_bus; > + > + vpci_bus = pci_scan_bus(VPCI_BUS_NUM, &vpci_ops, sysdata); > + if (vpci_bus) > + pr_err("create pci bus\n"); > + > + pci_bus_add_devices(vpci_bus); > + > + return 0; > +} > + > +/*==================== Virtual PCIe NTB driver ==========================*/ > + > +static int vntb_epf_mw_count(struct ntb_dev *ntb, int pidx) > +{ > + struct epf_ntb *ndev = ntb_ndev(ntb); > + > + return ndev->num_mws; > +} > + > +static int vntb_epf_spad_count(struct ntb_dev *ntb) > +{ > + return ntb_ndev(ntb)->spad_count; > +} > + > +static int vntb_epf_peer_mw_count(struct ntb_dev *ntb) > +{ > + return ntb_ndev(ntb)->num_mws; > +} > + > +static u64 vntb_epf_db_valid_mask(struct ntb_dev *ntb) > +{ > + return BIT_ULL(ntb_ndev(ntb)->db_count) - 1; > +} > + > +static int vntb_epf_db_set_mask(struct ntb_dev *ntb, u64 db_bits) > +{ > + return 0; > +} > + > +static int vntb_epf_mw_set_trans(struct ntb_dev *ndev, int pidx, int idx, > + dma_addr_t addr, resource_size_t size) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct pci_epf_bar *epf_bar; > + enum pci_barno barno; > + int ret; > + struct device *dev; > + > + dev = &ntb->ntb.dev; > + barno = ntb->epf_ntb_bar[BAR_MW0 + idx]; > + epf_bar = &ntb->epf->bar[barno]; > + epf_bar->phys_addr = addr; > + epf_bar->barno = barno; > + epf_bar->size = size; > + > + ret = pci_epc_set_bar(ntb->epf->epc, 0, 0, epf_bar); > + if (ret) { > + dev_err(dev, "failure set mw trans\n"); > + return ret; > + } > + return 0; > +} > + > +static int vntb_epf_mw_clear_trans(struct ntb_dev *ntb, int pidx, int idx) > +{ > + return 0; > +} > + > +static int vntb_epf_peer_mw_get_addr(struct ntb_dev *ndev, int idx, > + phys_addr_t *base, resource_size_t *size) > +{ > + > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + if (base) > + *base = ntb->vpci_mw_phy[idx]; > + > + if (size) > + *size = ntb->mws_size[idx]; > + > + return 0; > +} > + > +static int vntb_epf_link_enable(struct ntb_dev *ntb, > + enum ntb_speed max_speed, > + enum ntb_width max_width) > +{ > + return 0; > +} > + > +static u32 vntb_epf_spad_read(struct ntb_dev *ndev, int idx) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + int off = ntb->reg->spad_offset, ct = ntb->reg->spad_count * 4; > + u32 val; > + void __iomem *base = ntb->reg; > + > + val = readl(base + off + ct + idx * 4); > + return val; > +} > + > +static int vntb_epf_spad_write(struct ntb_dev *ndev, int idx, u32 val) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct epf_ntb_ctrl *ctrl = ntb->reg; > + int off = ctrl->spad_offset, ct = ctrl->spad_count * 4; > + void __iomem *base = ntb->reg; > + > + writel(val, base + off + ct + idx * 4); > + return 0; > +} > + > +static u32 vntb_epf_peer_spad_read(struct ntb_dev *ndev, int pidx, int idx) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct epf_ntb_ctrl *ctrl = ntb->reg; > + int off = ctrl->spad_offset; > + void __iomem *base = ntb->reg; > + u32 val; > + > + val = readl(base + off + idx * 4); > + return val; > +} > + > +static int vntb_epf_peer_spad_write(struct ntb_dev *ndev, int pidx, int idx, u32 val) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + struct epf_ntb_ctrl *ctrl = ntb->reg; > + int off = ctrl->spad_offset; > + void __iomem *base = ntb->reg; > + > + writel(val, base + off + idx * 4); > + return 0; > +} > + > +static int vntb_epf_peer_db_set(struct ntb_dev *ndev, u64 db_bits) > +{ > + u32 interrupt_num = ffs(db_bits) + 1; > + struct epf_ntb *ntb = ntb_ndev(ndev); > + u8 func_no, vfunc_no; > + int ret; > + > + func_no = ntb->epf->func_no; > + vfunc_no = ntb->epf->vfunc_no; > + > + ret = pci_epc_raise_irq(ntb->epf->epc, > + func_no, > + vfunc_no, > + PCI_EPC_IRQ_MSI, > + interrupt_num + 1); > + if (ret) { > + dev_err(&ntb->ntb.dev, "intf: Failed to raise IRQ\n"); > + return ret; > + } > + > + return 0; Equivalent to: if (ret) dev_err(...); return ret; > +} > + > +static u64 vntb_epf_db_read(struct ntb_dev *ndev) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + return ntb->db; > +} > + > +static int vntb_epf_mw_get_align(struct ntb_dev *ndev, int pidx, int idx, > + resource_size_t *addr_align, > + resource_size_t *size_align, > + resource_size_t *size_max) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + if (addr_align) > + *addr_align = SZ_4K; > + > + if (size_align) > + *size_align = 1; > + > + if (size_max) > + *size_max = ntb->mws_size[idx]; > + > + return 0; > +} > + > +static u64 vntb_epf_link_is_up(struct ntb_dev *ndev, > + enum ntb_speed *speed, > + enum ntb_width *width) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + return ntb->reg->link_status; > +} > + > +static int vntb_epf_db_clear_mask(struct ntb_dev *ndev, u64 db_bits) > +{ > + return 0; > +} > + > +static int vntb_epf_db_clear(struct ntb_dev *ndev, u64 db_bits) > +{ > + struct epf_ntb *ntb = ntb_ndev(ndev); > + > + ntb->db &= ~db_bits; > + return 0; > +} > + > +static int vntb_epf_link_disable(struct ntb_dev *ntb) > +{ > + return 0; > +} > + > +static const struct ntb_dev_ops vntb_epf_ops = { > + .mw_count = vntb_epf_mw_count, > + .spad_count = vntb_epf_spad_count, > + .peer_mw_count = vntb_epf_peer_mw_count, > + .db_valid_mask = vntb_epf_db_valid_mask, > + .db_set_mask = vntb_epf_db_set_mask, > + .mw_set_trans = vntb_epf_mw_set_trans, > + .mw_clear_trans = vntb_epf_mw_clear_trans, > + .peer_mw_get_addr = vntb_epf_peer_mw_get_addr, > + .link_enable = vntb_epf_link_enable, > + .spad_read = vntb_epf_spad_read, > + .spad_write = vntb_epf_spad_write, > + .peer_spad_read = vntb_epf_peer_spad_read, > + .peer_spad_write = vntb_epf_peer_spad_write, > + .peer_db_set = vntb_epf_peer_db_set, > + .db_read = vntb_epf_db_read, > + .mw_get_align = vntb_epf_mw_get_align, > + .link_is_up = vntb_epf_link_is_up, > + .db_clear_mask = vntb_epf_db_clear_mask, > + .db_clear = vntb_epf_db_clear, > + .link_disable = vntb_epf_link_disable, > +}; > + > +static int pci_vntb_probe(struct pci_dev *pdev, const struct pci_device_id *id) > +{ > + int ret; > + struct epf_ntb *ndev = (struct epf_ntb *)pdev->sysdata; > + struct device *dev = &pdev->dev; > + > + ndev->ntb.pdev = pdev; > + ndev->ntb.topo = NTB_TOPO_NONE; > + ndev->ntb.ops = &vntb_epf_ops; > + > + ret = dma_set_mask_and_coherent(dev, DMA_BIT_MASK(32)); > + if (ret) { > + dev_err(dev, "Cannot set DMA mask\n"); > + return -1; return -EINVAL; > + } > + > + ret = ntb_register_device(&ndev->ntb); > + if (ret) { > + dev_err(dev, "Failed to register NTB device\n"); > + goto err_register_dev; > + } > + > + dev_info(dev, "PCI Virtual NTB driver loaded\n"); > + return 0; > + > +err_register_dev: > + return -1; return -EINVAL; > +} > + > +static const struct pci_device_id pci_vntb_table[] = { > + { > + PCI_DEVICE(VNTB_VID, VNTB_PID), > + }, > + {}, > +}; > + > +static struct pci_driver vntb_pci_driver = { > + .name = "pci-vntb", > + .id_table = pci_vntb_table, > + .probe = pci_vntb_probe, > +}; > + > +/* ============ PCIe EPF Driver Bind ====================*/ > + > +/** > + * epf_ntb_bind() - Initialize endpoint controller to provide NTB functionality > + * @epf: NTB endpoint function device > + * > + * Initialize both the endpoint controllers associated with NTB function device. > + * Invoked when a primary interface or secondary interface is bound to EPC > + * device. This function will succeed only when EPC is bound to both the > + * interfaces. > + */ > +static int epf_ntb_bind(struct pci_epf *epf) > +{ > + struct epf_ntb *ntb = epf_get_drvdata(epf); > + struct device *dev = &epf->dev; > + int ret; > + > + if (!epf->epc) { > + dev_dbg(dev, "PRIMARY EPC interface not yet bound\n"); > + return 0; > + } > + > + ret = epf_ntb_init_epc_bar(ntb); > + if (ret) { > + dev_err(dev, "Failed to create NTB EPC\n"); > + goto err_bar_init; > + } > + > + ret = epf_ntb_config_spad_bar_alloc(ntb); > + if (ret) { > + dev_err(dev, "Failed to allocate BAR memory\n"); > + goto err_bar_alloc; > + } > + > + ret = epf_ntb_epc_init(ntb); > + if (ret) { > + dev_err(dev, "Failed to initialize EPC\n"); > + goto err_bar_alloc; > + } > + > + epf_set_drvdata(epf, ntb); > + > + if (pci_register_driver(&vntb_pci_driver)) { > + dev_err(dev, "failure register vntb pci driver\n"); > + goto err_bar_alloc; > + } > + > + vpci_bus(ntb); > + > + return 0; > + > +err_bar_alloc: > + epf_ntb_config_spad_bar_free(ntb); > + > +err_bar_init: > + epf_ntb_epc_destroy(ntb); > + > + return ret; > +} > + > +/** > + * epf_ntb_unbind() - Cleanup the initialization from epf_ntb_bind() > + * @epf: NTB endpoint function device > + * > + * Cleanup the initialization from epf_ntb_bind() > + */ > +static void epf_ntb_unbind(struct pci_epf *epf) > +{ > + struct epf_ntb *ntb = epf_get_drvdata(epf); > + > + epf_ntb_epc_cleanup(ntb); > + epf_ntb_config_spad_bar_free(ntb); > + epf_ntb_epc_destroy(ntb); > + > + pci_unregister_driver(&vntb_pci_driver); > +} > + > +// EPF driver probe > +static struct pci_epf_ops epf_ntb_ops = { > + .bind = epf_ntb_bind, > + .unbind = epf_ntb_unbind, > + .add_cfs = epf_ntb_add_cfs, > +}; > + > +/** > + * epf_ntb_probe() - Probe NTB function driver > + * @epf: NTB endpoint function device > + * > + * Probe NTB function driver when endpoint function bus detects a NTB > + * endpoint function. > + */ > +static int epf_ntb_probe(struct pci_epf *epf) > +{ > + struct epf_ntb *ntb; > + struct device *dev; > + > + dev = &epf->dev; > + > + ntb = devm_kzalloc(dev, sizeof(*ntb), GFP_KERNEL); > + if (!ntb) > + return -ENOMEM; > + > + epf->header = &epf_ntb_header; > + ntb->epf = epf; > + epf_set_drvdata(epf, ntb); > + > + dev_info(dev, "pci-ep epf driver loaded\n"); I think most drivers don't announce when they're loaded because it really doesn't help the user. > + return 0; > +} > + > +static const struct pci_epf_device_id epf_ntb_ids[] = { > + { > + .name = "pci_epf_vntb", > + }, > + {}, > +}; > + > +static struct pci_epf_driver epf_ntb_driver = { > + .driver.name = "pci_epf_vntb", > + .probe = epf_ntb_probe, > + .id_table = epf_ntb_ids, > + .ops = &epf_ntb_ops, > + .owner = THIS_MODULE, > +}; > + > + Spurious blank line. > +static int __init epf_ntb_init(void) > +{ > + int ret; > + > + kpcintb_workqueue = alloc_workqueue("kpcintb", WQ_MEM_RECLAIM | > + WQ_HIGHPRI, 0); > + ret = pci_epf_register_driver(&epf_ntb_driver); > + if (ret) { > + destroy_workqueue(kpcintb_workqueue); > + pr_err("Failed to register pci epf ntb driver --> %d\n", ret); > + return ret; > + } > + > + return 0; > +} > +module_init(epf_ntb_init); > + > +static void __exit epf_ntb_exit(void) > +{ > + pci_epf_unregister_driver(&epf_ntb_driver); > + destroy_workqueue(kpcintb_workqueue); > +} > +module_exit(epf_ntb_exit); > + > +MODULE_DESCRIPTION("PCI EPF NTB DRIVER"); > +MODULE_AUTHOR("Frank Li <Frank.li@xxxxxxx>"); > +MODULE_LICENSE("GPL v2"); > -- > 2.24.0.rc1 >