On Tue, Jan 19, 2016 at 11:58:06AM +0000, Singhal, Maneesh wrote: > Hello, > Kindly review the following patch for the following driver to be added in SCSI subsystem - > > Regards > Maneesh > Hi Maneesh, Fery first round of review. No real functionallity yet just a bit on readablility. I'll do a more in depth review later. > ---------------------------------------------------------------------------- > From f3c4b836d6f130b1d7ded618002c8164f8f4a06d Mon Sep 17 00:00:00 2001 > From: "maneesh.singhal" <maneesh.singhal@xxxxxxx> > Date: Tue, 19 Jan 2016 06:39:35 -0500 > Subject: [PATCH] [PATCH] drivers/scsi/emcctd: Client driver implementation for > EMC-Symmetrix GuestOS emulated Cut-Through Device. > > The patch is a driver implementation EMC-Symmetrix GuestOS emulated Cut-Through > Device. The Cut-Through Device PCI emulation is implemented for GuestOS > environments in the HyperMax OS. GuestOS environments allows loading of > any x86 compliant operating systems like Linux/FreeBSD etc. > > The client driver is a SCSI HBA implementation which interfaces with SCSI > midlayer in the north-bound interfaces and connects with the emulated PCI device > on the south side. > > The PCI vendor ID:product ID for emulated Cut-Through Device is 0x1120:0x1B00. > > Signed-off-by: maneesh.singhal <maneesh.singhal@xxxxxxx> > --- > Documentation/scsi/emcctd.txt | 57 + > MAINTAINERS | 9 + > drivers/scsi/Kconfig | 1 + > drivers/scsi/Makefile | 1 + > drivers/scsi/emcctd/Kconfig | 7 + > drivers/scsi/emcctd/Makefile | 1 + > drivers/scsi/emcctd/README | 10 + > drivers/scsi/emcctd/emc_ctd_interface.h | 386 +++++ > drivers/scsi/emcctd/emcctd.c | 2840 +++++++++++++++++++++++++++++++ > drivers/scsi/emcctd/emcctd.h | 232 +++ > 10 files changed, 3544 insertions(+) > create mode 100644 Documentation/scsi/emcctd.txt > create mode 100644 drivers/scsi/emcctd/Kconfig > create mode 100644 drivers/scsi/emcctd/Makefile > create mode 100644 drivers/scsi/emcctd/README > create mode 100644 drivers/scsi/emcctd/emc_ctd_interface.h > create mode 100644 drivers/scsi/emcctd/emcctd.c > create mode 100644 drivers/scsi/emcctd/emcctd.h > > diff --git a/Documentation/scsi/emcctd.txt b/Documentation/scsi/emcctd.txt > new file mode 100644 > index 0000000..bcafc87 > --- /dev/null > +++ b/Documentation/scsi/emcctd.txt > @@ -0,0 +1,56 @@ > +This file contains brief information about the EMC Cut-Through Driver (emcctd). > +The driver is currently maintained by Singhal, Maneesh (maneesh.singhal@xxxxxxx) > + > +Last modified: Mon Jan 18 2016 by Maneesh Singhal > + > +BASICS > + > +Its a client driver implementation for EMC-Symmetrix GuestOS emulated > +Cut-Through Device. The Cut-Through Device PCI emulation is implemented for > +GuestOS environments in the HyperMax OS. GuestOS environments allows loading of > +any x86 compliant operating systems like Linux/FreeBSD etc. > + > +The client driver is a SCSI HBA implementation which interfaces with SCSI > +midlayer in the north-bound interfaces and connects with the emulated PCI device > +on the south side. > + > +The PCI vendor ID:product ID for emulated Cut-Through Device is 0x1120:0x1B00. > + > +VERSIONING > + > +The Version of the driver is maintained as 2.0.0.X, where 2 refers to the CTD > +protocol in use, and X refers to the ongoing version of the driver. > + > + > +SYSFS SUPPORT > + > +The driver creates the directory /sys/module/emcctd and populates it with > +version file and a directory for various parameters as described in MODULE > +PARAMETERS section. > + > +PROCFS SUPPORT > + > +The driver creates the directory /proc/emc and creates files emcctd_stats_x > +where 'x' refers to the PCI emulation number this client driver connected to. > +These files cotains WWN information and IO statistics for the particular PCI > +emulation. > + > +MODULE PARAMETERS > + > +The supported parameters which could add debuggability or change the runtime > +behavior of the driver are as following: > + > +ctd_debug=0 | 1 Enable driver debug messages(0=off, 1=on) > + > +max_luns=xx Specify the maximum number of LUN's per > + host(default=16384) > + > +cmd_per_lun=xx Specify the maximum commands per lun(default=16) > + > +DEBUGGING HINTS > + > +Debugging code is now compiled in by default but debugging is turned off > +with the kernel module parameter debug_flag defaulting to 0. To enable debug at > +module load time add ctd_debug=1 to the module load options. > +Debugging can also be enabled and disabled by writing a '0' (disable) or '1' > +(enable) to the sysfs file /sys/module/emcctd/parameters/ctd_debug > diff --git a/MAINTAINERS b/MAINTAINERS > index 1d23f70..671173a 100644 > --- a/MAINTAINERS > +++ b/MAINTAINERS > @@ -6601,6 +6601,15 @@ F: drivers/message/fusion/ > F: drivers/scsi/mpt2sas/ > F: drivers/scsi/mpt3sas/ > > +EMC Cut-Through Driver > +M: fredette, matt <matt.fredette@xxxxxxx> > +M: Pirotte, Serge <serge.pirotte@xxxxxxx> > +M: Singh Animesh <Animesh.Singh@xxxxxxx> > +M: Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > +L: linux-scsi@xxxxxxxxxxxxxxx > +S: Maintained > +F: drivers/scsis/emcctd/ > + > LSILOGIC/SYMBIOS/NCR 53C8XX and 53C1010 PCI-SCSI drivers > M: Matthew Wilcox <matthew@xxxxxx> > L: linux-scsi@xxxxxxxxxxxxxxx > diff --git a/drivers/scsi/Kconfig b/drivers/scsi/Kconfig > index c1fe0d2..cbc03d7 100644 > --- a/drivers/scsi/Kconfig > +++ b/drivers/scsi/Kconfig > @@ -477,6 +477,7 @@ source "drivers/scsi/aic7xxx/Kconfig.aic79xx" > source "drivers/scsi/aic94xx/Kconfig" > source "drivers/scsi/hisi_sas/Kconfig" > source "drivers/scsi/mvsas/Kconfig" > +source "drivers/scsi/emcctd/Kconfig" > > config SCSI_MVUMI > tristate "Marvell UMI driver" > diff --git a/drivers/scsi/Makefile b/drivers/scsi/Makefile > index 862ab4e..55bea65 100644 > --- a/drivers/scsi/Makefile > +++ b/drivers/scsi/Makefile > @@ -132,6 +132,7 @@ obj-$(CONFIG_SCSI_IBMVFC) += ibmvscsi/ > obj-$(CONFIG_SCSI_HPTIOP) += hptiop.o > obj-$(CONFIG_SCSI_STEX) += stex.o > obj-$(CONFIG_SCSI_MVSAS) += mvsas/ > +obj-$(CONFIG_SCSI_EMCCTD) += emcctd/ > obj-$(CONFIG_SCSI_MVUMI) += mvumi.o > obj-$(CONFIG_PS3_ROM) += ps3rom.o > obj-$(CONFIG_SCSI_CXGB3_ISCSI) += libiscsi.o libiscsi_tcp.o cxgbi/ > diff --git a/drivers/scsi/emcctd/Kconfig b/drivers/scsi/emcctd/Kconfig > new file mode 100644 > index 0000000..d995b53 > --- /dev/null > +++ b/drivers/scsi/emcctd/Kconfig > @@ -0,0 +1,7 @@ > +config SCSI_EMCCTD > + tristate "EMC Cut-Through client driver" > + depends on SCSI && PCI && X86 > + help > + This driver supports EMC virtual PCI Cut-Through-Device. > + To compile this driver as a module, choose M here: the > + module will be called emcctd. > diff --git a/drivers/scsi/emcctd/Makefile b/drivers/scsi/emcctd/Makefile > new file mode 100644 > index 0000000..74a070e > --- /dev/null > +++ b/drivers/scsi/emcctd/Makefile > @@ -0,0 +1 @@ > +obj-$(CONFIG_SCSI_EMCCTD) += emcctd.o > diff --git a/drivers/scsi/emcctd/README b/drivers/scsi/emcctd/README > new file mode 100644 > index 0000000..496ce7f > --- /dev/null > +++ b/drivers/scsi/emcctd/README > @@ -0,0 +1,10 @@ > +EMCCTD : Client driver implementation for EMC-Symmetrix GuestOS emulated > +Cut-Through Device. The Cut-Through Device PCI emulation is implemented for > +GuestOS environments in the HyperMax OS. GuestOS environments allows loading of > +any x86 compliant operating systems like Linux/FreeBSD etc. > + > +The client driver is a SCSI HBA implementation which interfaces with SCSI > +midlayer in the north-bound interfaces and connects with the emulated PCI device > +on the south side. > + > +The PCI vendor ID:product ID for emulated Cut-Through Device is 0x1120:0x1B00. Wouldn't it be nice to have this in the Kconfig file? No user will ever look at the README file in the driver directory. > diff --git a/drivers/scsi/emcctd/emc_ctd_interface.h b/drivers/scsi/emcctd/emc_ctd_interface.h > new file mode 100644 > index 0000000..58a0276 > --- /dev/null > +++ b/drivers/scsi/emcctd/emc_ctd_interface.h > @@ -0,0 +1,386 @@ > +/* > + * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem. > + * > + * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA. > + * > + * Authors: > + * fredette, matt <matt.fredette@xxxxxxx> > + * Pirotte, Serge <serge.pirotte@xxxxxxx> > + * Singh Animesh <Animesh.Singh@xxxxxxx> > + * Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > + * > + * This program is free software; you can redistribute it and/or > + * modify it under the terms of the GNU General Public License > + * as published by the Free Software Foundation; either version 2 > + * of the License, or (at your option) any later version. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + * > + */ > + > +#ifndef _EMC_CTD_INTERFACE_H > +#define _EMC_CTD_INTERFACE_H > +/* > + * > + * DESCRIPTION: Data structures used by the Cut-through subsystem. > + * > + * NOTES: Changes to any of these structures will mean that any Clients that > + * depend on them will also need to be modified. Since many of those > + * Clients are not part of the Enginuity build process, this will almost > + * certainly require new versions of Guest OS CTD Client code to be > + * released. > + * > + * Before modifying any data structures in this file please discuss the > + * change with the maintainers. > + * > + */ > + > + > +/* macros: */ > + > +/* the PCI vendor ID for all devices: */ > +#define EMC_CTD_PCI_VENDOR (0x1120) > + > +/* the PCI product ID for all version 1.x devices: */ > +#define EMC_CTD_V010_PCI_PRODUCT (0x1b00) > + > +/* the PCI revision ID for the first version 1.0 device: */ > +#define EMC_CTD_V010_PCI_REVISION (1) > + > +/* the 64-bit BAR pair for the transmit and receive rings: */ > +#define EMC_CTD_V010_BAR_RINGS (0) > + > +/* the 64-bit BAR pair for the fast registers: */ > +#define EMC_CTD_V010_BAR_FREGS (2) > + > +/* the 32-bit BAR for the slow registers: */ > +#define EMC_CTD_V010_BAR_SREGS (4) > + > +/* the maximum number of immediate SGL entries: */ > +#define EMC_CTD_V010_SGL_IMMEDIATE_MAX (7) > + > +/* the CTD v010 whats: */ > +#define EMC_CTD_V010_WHAT_DETECT (0) > +#define EMC_CTD_V010_WHAT_SCSI_COMMAND (1) > +#define EMC_CTD_V010_WHAT_SCSI_PHASE (2) > +#define EMC_CTD_V010_WHAT_SCSI_RESPONSE (3) > + > +/* the CTD v010 detect flags. all undefined flags must be zero: */ > + > +/* if this is set, the name is a SCSI target: */ > +#define EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET (1 << 0) > + > +/* if this is set, the name is a SCSI initiator: */ > +#define EMC_CTD_V010_DETECT_FLAG_SCSI_INITIATOR (1 << 1) > + > +/* the CTD v010 SCSI command flags. all undefined flags must be zero: */ > + > +/* when the guest receives a SCSI command message, this flag is undefined. > + * > + * If this is set, at the beginning of any data phase the target is the data > + * source. if this is clear, at the beginning of any data phase the target is > + * the data sink. > + */ > +#define EMC_CTD_V010_SCSI_COMMAND_FLAG_SOURCE (1 << 0) > + > +/* when the guest receives a SCSI command message, this flag is undefined. > + * > + * if this is set, the first SGL entry in the message points to an extended SGL, > + * and the remaining SGL entries in the message are undefined. if this is > + * clear, the SGL entries in the message are used: > + */ > +#define EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL (1 << 1) > + > +/* the CTD v010 SCSI response flags. all undefined flags must be zero: */ > + > +/* if this is set, the SCSI command failed. if this is clear, the > + * command succeeded: > + */ > +#define EMC_CTD_V010_SCSI_RESPONSE_FLAG_FAILED (1 << 0) > + > +/* if this is set, any extra information is sense data. if this is clear, any > + * extra information is 64-bit timestamps: > + */ > +#define EMC_CTD_V010_SCSI_RESPONSE_FLAG_SENSE (1 << 1) > + > +/* the CTD v010 SCSI phase flags. all undefined flags must be zero: */ > + > +/* when the guest receives a SCSI phase message, this flag is undefined. > + * > + * if this is set, at this point in the data phase the message receiver is the > + * data source. if this is clear, at this point in the data phase the message > + * receiver is the data sink: > + */ > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_SOURCE (1 << 0) > + > +/* when the guest receives a SCSI phase message, this flag is undefined. > + * > + * if this is set, the first SGL entry in the message points to an extended SGL, > + * and the remaining SGL entries in the message are undefined. if this is > + * clear, the SGL entries in the message are used: > + */ > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_ESGL (1 << 1) > + > +/* if this is set, the message receiver is the target. if this is clear, the > + * message receiver is the initiator: > + */ > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET (1 << 2) > + > +/* if this is set, the SCSI command is aborted: */ > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT (1 << 3) > + Please use the BIT() macro for the above bit definitions. > +/* the size of the log of errored transmit messages: */ > +#define EMC_CTD_V010_LOG_ERROR_TX_SIZE (4) > + > +/* errors: */ > + > +/* no error: */ > +#define EMC_CTD_V010_ERROR_NULL (0) > + > +/* the guest tried to transmit a message on a disconnected channel: */ > +#define EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED (1) > + > +/* the guest tried to transmit a message with a bad what: */ > +#define EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT (2) > + > +/* the guest tried to transmit a message with a reserved field set to > + * the wrong value: > + */ > +#define EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED (3) > + > +/* the guest tried to transmit an out-of-order message: */ > +#define EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER (4) > + > +/* the guest tried to transmit a message to an endpoint whose type > + * doesn't support it: > + */ > +#define EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE (5) > + > +/* the guest tried to transmit a message with an unknown message > + * receiver's opaque value: > + */ > +#define EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN (6) > + > +/* types: */ > + > +/* a CTD v010 scatter/gather list entry: */ > +struct emc_ctd_v010_sgl { > + > + /* the physical address of the buffer: */ > + emc_ctd_uint32_t emc_ctd_v010_sgl_paddr_0_31; > + emc_ctd_uint32_t emc_ctd_v010_sgl_paddr_32_63; > + > + /* the size of the buffer: */ > + emc_ctd_uint32_t emc_ctd_v010_sgl_size; > +}; > + > +/* a CTD v010 header: */ > +struct emc_ctd_v010_header { > + > + /* the other address: */ > + emc_ctd_uint16_t emc_ctd_v010_header_address; > + > + /* the minor version: */ > + emc_ctd_uint8_t emc_ctd_v010_header_minor; > + > + /* the what: */ > + emc_ctd_uint8_t emc_ctd_v010_header_what; > +}; Well this is a matter of taste but you have (and not only in this struct, just an example) emc_ctd_v010_header.emc_ctd_v010_header_address all the emc_ctd_v010_header_ stuff is totally redundant and you suffer from extremely long lines in your dirver anyways. Just a hint. > + > +/* a CTD v010 name: */ > +struct emc_ctd_v010_name { > + > + /* the name: */ > + emc_ctd_uint8_t emc_ctd_v010_name_bytes[8]; > +}; > + > +/* a CTD v010 detect message: */ > +struct emc_ctd_v010_detect { > + > + /* the header: */ > + struct emc_ctd_v010_header emc_ctd_v010_detect_header; > + > + /* the flags: */ > + emc_ctd_uint32_t emc_ctd_v010_detect_flags; > + > + /* the name: */ > + struct emc_ctd_v010_name emc_ctd_v010_detect_name; > + > + /* the key: */ > + emc_ctd_uint64_t emc_ctd_v010_detect_key; > +}; > + > +/* a CTD v010 SCSI command message: */ > +struct emc_ctd_v010_scsi_command { > + > + /* the header: */ > + struct emc_ctd_v010_header emc_ctd_v010_scsi_command_header; > + > + /* the flags: */ > + emc_ctd_uint32_t emc_ctd_v010_scsi_command_flags; > + > + /* the initiator's opaque value: */ > + emc_ctd_uint64_t emc_ctd_v010_scsi_command_opaque; > + > + /* the SCSI LUN: */ > + emc_ctd_uint8_t emc_ctd_v010_scsi_command_lun[8]; > + > + /* the SCSI CDB: */ > + emc_ctd_uint8_t emc_ctd_v010_scsi_command_cdb[16]; > + > + /* the data size: */ > + emc_ctd_uint32_t emc_ctd_v010_scsi_command_data_size; > + > + union { > + > + /* any SGL entries: */ > + /* when received by the guest, these are undefined: */ > + struct emc_ctd_v010_sgl > + emc_ctd_v010_scsi_command_sgl[EMC_CTD_V010_SGL_IMMEDIATE_MAX]; > + > + } emc_ctd_v010_scsi_command_u; > +}; > + > +/* a CTD v010 SCSI response message: */ > +struct emc_ctd_v010_scsi_response { > + > + /* the header: */ > + struct emc_ctd_v010_header emc_ctd_v010_scsi_response_header; > + > + /* the flags: */ > + emc_ctd_uint16_t emc_ctd_v010_scsi_response_flags; > + > + /* the extra information size: */ > + emc_ctd_uint8_t emc_ctd_v010_scsi_response_extra_size; > + > + /* the SCSI status: */ > + emc_ctd_uint8_t emc_ctd_v010_scsi_response_status; > + > + /* the initiator's opaque value: */ > + emc_ctd_uint64_t emc_ctd_v010_scsi_response_opaque; > + > + /* the data size: */ > + emc_ctd_uint32_t emc_ctd_v010_scsi_response_data_size; > + > + union { > + > + /* any extra information: */ > + emc_ctd_uint8_t emc_ctd_v010_scsi_response_extra[108]; > + > + } emc_ctd_v010_scsi_response_u; > +}; > + > +/* a CTD v010 SCSI phase message: */ > +struct emc_ctd_v010_scsi_phase { > + > + /* the header: */ > + struct emc_ctd_v010_header emc_ctd_v010_scsi_phase_header; > + > + /* the flags: */ > + emc_ctd_uint32_t emc_ctd_v010_scsi_phase_flags; > + > + /* the message receiver's opaque value: */ > + emc_ctd_uint64_t emc_ctd_v010_scsi_phase_opaque_rx; > + > + /* the message transmitter's opaque value: */ > + emc_ctd_uint64_t emc_ctd_v010_scsi_phase_opaque_tx; > + > + union { > + > + /* any SGL entries: */ > + /* when received by the guest, these are undefined: */ > + struct emc_ctd_v010_sgl > + emc_ctd_v010_scsi_phase_sgl[EMC_CTD_V010_SGL_IMMEDIATE_MAX]; > + > + } emc_ctd_v010_scsi_phase_u; > +}; > + > +/* a CTD v010 message: */ > +union emc_ctd_v010_message { > + > + /* the header: */ > + struct emc_ctd_v010_header emc_ctd_v010_message_header; > + > + /* a detect message: */ > + struct emc_ctd_v010_detect emc_ctd_v010_message_detect; > + > + /* a SCSI command message: */ > + struct emc_ctd_v010_scsi_command emc_ctd_v010_message_scsi_command; > + > + /* a SCSI response message: */ > + struct emc_ctd_v010_scsi_response emc_ctd_v010_message_scsi_response; > + > + /* a SCSI phase message: */ > + struct emc_ctd_v010_scsi_phase emc_ctd_v010_message_scsi_phase; > + > + /* padding: */ > + emc_ctd_uint8_t emc_ctd_v010_message_padding[128]; > +}; > + > +/* the fast registers: */ > +struct emc_ctd_v010_fregs { > + > + /* the transmit ring producer index (TPI): */ > + volatile emc_ctd_uint32_t emc_ctd_v010_fregs_tx_index_producer; > + > + /* the error flag: */ > + volatile emc_ctd_uint32_t emc_ctd_v010_fregs_error_flag; > + > + /* errors 1..14: */ > + volatile emc_ctd_uint32_t emc_ctd_v010_fregs_errors_1_14[14]; > + > + /* the transmit ring consumer index (TCI): */ > + volatile emc_ctd_uint32_t emc_ctd_v010_fregs_tx_index_consumer; > + > + /* the device name: */ > + struct emc_ctd_v010_name emc_ctd_v010_fregs_device_name; > + > + /* padding: */ > + emc_ctd_uint8_t > + emc_ctd_v010_fregs_pad_07f[64 - > + (sizeof(emc_ctd_uint32_t) + > + sizeof(struct emc_ctd_v010_name))]; Maybe: #define PAD07F_SIZE (64 - sizeof(u32) + sizeof(struct emc_ctd_v010_name)) u8 emc_ctd_v010_fregs_pad_07f[PAD07f_SIZE]; > + > + /* the receive ring producer index (RPI): */ > + volatile emc_ctd_uint32_t emc_ctd_v010_fregs_rx_index_producer; > + > + /* the interrupt throttle, in units of nanoseconds. zero disables > + * the throttle: > + */ > + emc_ctd_uint32_t emc_ctd_v010_fregs_interrupt_throttle_nsecs; > + > + /* padding: */ > + emc_ctd_uint8_t > + emc_ctd_v010_fregs_pad_0bf[64 - > + (sizeof(emc_ctd_uint32_t) + sizeof(emc_ctd_uint32_t))]; Same applies here. > + > + /* the receive ring consumer index (RCI): */ > + volatile emc_ctd_uint32_t emc_ctd_v010_fregs_rx_index_consumer; > + > + /* padding: */ > + emc_ctd_uint8_t > + emc_ctd_v010_fregs_pad_0ff[64 - > + (sizeof(emc_ctd_uint32_t) + > + (sizeof(emc_ctd_uint32_t) * > + EMC_CTD_V010_LOG_ERROR_TX_SIZE))]; And here. > + > + /* the errors for the log of errored transmit messages: */ > + volatile emc_ctd_uint32_t > + emc_ctd_v010_fregs_log_error_tx_error[EMC_CTD_V010_LOG_ERROR_TX_SIZE]; > + > + /* the log of errored transmit messages: */ > + union emc_ctd_v010_message > + emc_ctd_v010_fregs_log_error_tx_message[EMC_CTD_V010_LOG_ERROR_TX_SIZE]; > +}; > + > +/* the slow registers: */ > +struct emc_ctd_v010_sregs { > + > + /* the reset register: */ > + emc_ctd_uint32_t emc_ctd_v010_sregs_reset; > +}; > + > +#endif /* _EMC_CTD_INTERFACE_H */ > diff --git a/drivers/scsi/emcctd/emcctd.c b/drivers/scsi/emcctd/emcctd.c > new file mode 100644 > index 0000000..aa6dd0e > --- /dev/null > +++ b/drivers/scsi/emcctd/emcctd.c > @@ -0,0 +1,2840 @@ > +/* > + * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem. > + * > + * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA. > + * > + * Authors: > + * fredette, matt <matt.fredette@xxxxxxx> > + * Pirotte, Serge <serge.pirotte@xxxxxxx> > + * Singh Animesh <Animesh.Singh@xxxxxxx> > + * Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > + * > + * This program is free software; you can redistribute it and/or > + * modify it under the terms of the GNU General Public License > + * as published by the Free Software Foundation; either version 2 > + * of the License, or (at your option) any later version. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + * > + */ > + > +#include <linux/init.h> > +#include <linux/kernel.h> > +#include <linux/kthread.h> > +#include <linux/slab.h> > +#include <linux/list.h> > +#include <linux/module.h> > +#include <linux/spinlock.h> > +#include <linux/pci.h> > +#include <linux/interrupt.h> > +#include <linux/proc_fs.h> > +#include <linux/version.h> > +#include <linux/delay.h> > +#include <linux/blkdev.h> > +#include <linux/atomic.h> > + > +#include <linux/cache.h> > + > +#include <linux/seq_file.h> > + > +#include <scsi/scsi.h> > +#include <scsi/scsi_cmnd.h> > +#include <scsi/scsi_device.h> > +#include <scsi/scsi_host.h> > +#include <scsi/scsi_transport.h> > +#include <scsi/scsi_dbg.h> > + > +#define emc_ctd_uint8_t u8 > +#define emc_ctd_uint16_t u16 > +#define emc_ctd_uint32_t u32 > +#define emc_ctd_uint64_t u64 Please don't define your own data types and use u8, u16, u32 and u64 directly. > + > +#include "emc_ctd_interface.h" > + > +#include "emcctd.h" > + > +#include <scsi/scsi_dbg.h> > + > +/* nomenclature for versioning > + * MAJOR:MINOR:SUBVERSION:PATCH > + */ > + > +#define EMCCTD_MODULE_VERSION "2.0.0.24" > + > +MODULE_LICENSE("GPL"); > +MODULE_AUTHOR("EMC"); > +MODULE_DESCRIPTION("EMC CTD V1 - Build 18-Jan-2016"); This is very misleading. If I was a user I'd think the kernel was build on 18-Jan-2016. Anyways The version should be enough. > +MODULE_VERSION(EMCCTD_MODULE_VERSION); > + > +int ctd_debug; > +module_param_named(ctd_debug, ctd_debug, int, S_IRUGO | S_IWUSR); > +MODULE_PARM_DESC(ctd_debug, "Enable driver debug messages(0=off, 1=on)"); > + > +static int emcctd_max_luns = EMCCTD_MAX_LUN; > +module_param_named(max_luns, emcctd_max_luns, int, S_IRUGO | S_IWUSR); > +MODULE_PARM_DESC(max_luns, > + "Specify the maximum number of LUN's per host(default=16384)"); > + > +static int emcctd_cmd_per_lun = EMCCTD_CMD_PER_LUN; > +module_param_named(cmd_per_lun, emcctd_cmd_per_lun, int, S_IRUGO | S_IWUSR); > +MODULE_PARM_DESC(cmd_per_lun, > + "Specify the maximum commands per lun(default=64)"); > + > +/* > + * routines to handle scsi cmd buffer > + */ > +static int ctd_xmit_command(struct scsi_cmnd *cmnd, > + struct ctd_pci_private *ctd_private); > +static void ctd_handle_scsi_response( > + struct emc_ctd_v010_scsi_response *io_response, > + struct ctd_pci_private *ctd_private); > + > +static int ctd_hw_enqueue_request(union emc_ctd_v010_message *ctd_request, > + struct ctd_pci_private *ctd_private); > +static int ctd_hw_dequeue_response(union emc_ctd_v010_message *ctd_response, > + struct ctd_pci_private *ctd_private); > +static int ctd_scsi_response_sanity_check( > + struct ctd_request_private *request_private, > + struct ctd_pci_private *ctd_private); > + > +/* > + * routines to handle memory and queue management between client and server > + */ > +static int ctd_hw_execute_command(struct scsi_cmnd *cmnd, > + struct ctd_pci_private *ctd_private); > +static int ctd_initiator_translate_sgl(struct scsi_cmnd *cmnd, > + struct emc_ctd_v010_scsi_command *ctd_request, > + struct ctd_pci_private *ctd_private); > +static int ctd_initiator_translate_request(struct scsi_cmnd *cmnd, > + struct emc_ctd_v010_scsi_command *ctd_request, > + struct ctd_pci_private *ctd_private); > +static void ctd_initiator_translate_lun(struct scsi_cmnd *cmnd, > + struct emc_ctd_v010_scsi_command *ctd_request); > + > +static struct ctd_request_private *ctd_acquire_request( > + struct ctd_pci_private *ctd_private); > +static void ctd_release_request(struct ctd_request_private *ctd_request, > + struct ctd_pci_private *ctd_private); > +static void ctd_release_io_pool(struct ctd_pci_private *ctd_private); > +static int ctd_alloc_io_pool(struct ctd_pci_private *ctd_private, > + unsigned int pool_size); > +static void ctd_check_error_condition(struct pci_dev *pci_dev); > +static int ctd_init_event_thread(struct ctd_pci_private *ctd_private); > +static void ctd_destroy_event_thread(struct ctd_pci_private *ctd_private); > +/* > + * routines related to tasklet functionality > + */ > +static void ctd_check_response_queue(unsigned long instance_addr); > + > +static int ctd_ITnexus_handler(struct ctd_pci_private *ctd_private); > +/* > + * routines related to internal representation of scsi targets > + */ > +static int ctd_target_alloc(struct scsi_target *starget); > +static void ctd_target_destroy(struct scsi_target *starget); > +/* > + * routines registered with the linux scsi midlayer > + */ > +static int ctd_queuecommand_lck(struct scsi_cmnd *cmnd, > + void (*done)(struct scsi_cmnd *)); > +static DEF_SCSI_QCMD(ctd_queuecommand) > +static int ctd_slave_configure(struct scsi_device *sdp); > +static int ctd_slave_alloc(struct scsi_device *sdev); > +static void ctd_slave_destroy(struct scsi_device *sdev); > +static int ctd_abort_handler(struct scsi_cmnd *cmd); > + > +static enum blk_eh_timer_return ctd_timeout_handler(struct scsi_cmnd *scmd); > + > + > +/* > + * routines related to initialization of the pseudo hardware > + */ > +static void ctd_init_scsi_host_private(struct Scsi_Host *shost, > + struct pci_dev *pci_dev); > + > +static int ctd_scsi_layer_init(struct pci_dev *pci_dev); > +static int ctd_scsi_layer_cleanup(struct pci_dev *pci_dev); > + > +#ifdef CONFIG_PM > +static int ctd_pci_suspend(struct pci_dev *pci_dev, pm_message_t state); > +static int ctd_pci_resume(struct pci_dev *pci_dev); > +#endif > + > +static void ctd_pci_remove(struct pci_dev *pci_dev); > +static int ctd_request_msi(struct pci_dev *pci_dev); > +static int ctd_pci_probe(struct pci_dev *pci_dev, > + const struct pci_device_id *id); > +static void ctd_clear_io_queue(struct ctd_pci_private *ctd_private); > + > +#if !defined(__VMKLNX__) This is not VMWare's kernel but the upstream kernel, so please don't use that ifdefs here. > +static int ctd_proc_init(struct pci_dev *pci_dev); > +static void ctd_proc_remove(struct pci_dev *pci_dev); > + > +static int ctd_proc_open(struct inode *inode, struct file *file); > +#endif > + > +static int ctd_post_event(union emc_ctd_v010_message *io_msg, > + struct ctd_pci_private *ctd_private); > +static int ctd_event_handler(void *ctd_thread_args); > + > +unsigned int lun_discovery_complete; > +wait_queue_head_t lun_discovery_event_barrier; > + > +static const struct pci_device_id ctd_pci_id_table[] = { > + { EMC_CTD_PCI_VENDOR, EMC_CTD_V010_PCI_PRODUCT, > + PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 }, > + { 0 }, > +}; > + > +MODULE_DEVICE_TABLE(pci, ctd_pci_id_table); > + > +static struct pci_driver ctd_pci_driver = { > + .name = "emcctd", > + .id_table = ctd_pci_id_table, > + .probe = ctd_pci_probe, > + .remove = ctd_pci_remove, > +#ifdef CONFIG_PM > + .suspend = ctd_pci_suspend, > + .resume = ctd_pci_resume, > +#endif > +}; > + > +static struct scsi_host_template scsi_ctd_template = { > + .name = DRV_NAME, > + .proc_name = DRV_NAME, > + .queuecommand = ctd_queuecommand, > + > + .eh_timed_out = ctd_timeout_handler, > + > + .slave_alloc = ctd_slave_alloc, > + .slave_configure = ctd_slave_configure, > + .slave_destroy = ctd_slave_destroy, > + .eh_abort_handler = ctd_abort_handler, > + .target_alloc = ctd_target_alloc, > + .target_destroy = ctd_target_destroy, > + .can_queue = EMCCTD_CMD_PER_LUN, > + .this_id = EMCCTD_THIS_ID, > + .sg_tablesize = SG_ALL, > + .max_sectors = SCSI_DEFAULT_MAX_SECTORS, > + .cmd_per_lun = EMCCTD_CMD_PER_LUN, > + .use_clustering = DISABLE_CLUSTERING, > + .module = THIS_MODULE, > +}; > + > + No proc interface please. > +#if !defined(__VMKLNX__) > +static struct proc_dir_entry *ctd_proc_directory; > + > +static int > +ctd_proc_show(struct seq_file *m, void *v) > +{ > + int i; > + struct ctd_pci_private *ctd_private = > + (struct ctd_pci_private *)m->private; > + > + seq_printf(m, > + "number interrupts: %ld\n" > + "requests queued: %ld\n" > + "responses received: %ld\n" > + "pending IO count: %ld\n" > + "Abort Sent: %ld\n" > + "Abort received: %ld\n" > + "What received: %ld\n" > + "What sent: %ld\n" > + "free IO entries : %ld\n" > + "CTD WWN: %x.%x.%x.%x.%x.%x.%x.%x\n", > + ctd_private->hw_stats.interrupts.counter, > + ctd_private->hw_stats.requests_sent.counter, > + ctd_private->hw_stats.responses_received.counter, > + ctd_private->hw_stats.active_io_count.counter, > + ctd_private->hw_stats.abort_sent.counter, > + ctd_private->hw_stats.abort_received.counter, > + ctd_private->hw_stats.what_in.counter, > + ctd_private->hw_stats.what_out.counter, > + ctd_private->hw_stats.free_io_entries.counter, > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[0], > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[1], > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[2], > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[3], > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[4], > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[5], > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[6], > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[7]); > + > +#define MAX_ENTRIES_IN_LINE 10 > + seq_puts(m, "\nIO Latency (in tsc) for last 200 IOs:\n"); > + > + for (i = 0; i < CTD_MAX_IO_STATS; i++) { > + if (0 == (i % MAX_ENTRIES_IN_LINE)) > + seq_puts(m, "\n"); > + > + seq_printf(m, "%lld \t", ctd_private->hw_stats.io_stats[i]); > + } > + seq_puts(m, "\n"); > + > + return 0; > + > +} > + > +#endif > + > +static void > +scsi_translate_sam_code(struct scsi_cmnd *cmnd, unsigned char ScsiStatus) > +{ > + unsigned char host_status; > + unsigned char driver_status; > + > + host_status = DID_OK; > + driver_status = DRIVER_OK; > + cmnd->result |= ScsiStatus & 0xff; > + > + switch (ScsiStatus) { > + > + case SAM_STAT_GOOD: > + case SAM_STAT_CONDITION_MET: > + case SAM_STAT_INTERMEDIATE_CONDITION_MET: > + case SAM_STAT_INTERMEDIATE: > + break; > + case SAM_STAT_CHECK_CONDITION: > + case SAM_STAT_RESERVATION_CONFLICT: > + case SAM_STAT_ACA_ACTIVE: > + driver_status = DRIVER_SENSE; > + break; > + case SAM_STAT_TASK_SET_FULL: > + case SAM_STAT_BUSY: > + driver_status = DRIVER_BUSY; > + host_status = DID_REQUEUE; > + break; > + case SAM_STAT_TASK_ABORTED: > + driver_status = DRIVER_ERROR; > + host_status = DID_ABORT; > + break; > + case SAM_STAT_COMMAND_TERMINATED: > + default: > + ctd_dprintk_crit( > + "cmnd = %p [ channel:%d id:%d lun:%lld] INVALID SAM = %x\n", > + cmnd, cmnd->device->channel, cmnd->device->id, > + cmnd->device->lun, ScsiStatus); > + > + driver_status = DRIVER_INVALID; > + host_status = DID_ABORT; > + break; > + } > + set_driver_byte(cmnd, driver_status); > + set_host_byte(cmnd, host_status); > +} > + > +static void > +scsi_free_ctd_request_private(struct ctd_request_private *request_private, > + struct ctd_pci_private *ctd_private) > +{ > + > + if (request_private->cdb_page) { > + __free_pages(request_private->cdb_page, > + request_private->cdb_page_order); > + } > + if (request_private->sgllist_page) { > + __free_pages(request_private->sgllist_page, > + request_private->sgllist_page_order); > + } > + > + ctd_release_request(request_private, ctd_private); > +} > + > +static int > +ctd_handle_disconnect(struct emc_ctd_v010_detect *io_detect, > + struct ctd_pci_private *ctd_private) > +{ > + int i; > + int error; > + struct ctd_target_info *ctd_target; > + struct ctd_host_info *ctd_host; > + > + error = -ENODEV; > + > + ctd_dprintk_crit("\n"); > + > + ctd_host = ctd_private->host_private; > + > + /* Current implementation only handles the disconnect of > + * the target and not initiators > + */ > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > + ctd_target = &ctd_host->target[i]; > + > + /* detect header address is used to uniquely identify the target > + * for which the disconnect event has been posted by the server > + */ > + > + if (ctd_target->ctd_detect.emc_ctd_detect_header_address == > + io_detect->emc_ctd_detect_header_address) { > + /* check the current link status of the target */ > + if (ctd_target->ctd_detect.emc_ctd_v010_detect_flags) { > + ctd_dprintk_crit("\n"); > + > + ctd_target->ctd_detect.emc_ctd_v010_detect_flags = > + io_detect->emc_ctd_v010_detect_flags; > + > + ctd_check_error_condition(ctd_private->pci_dev); > + } else { > + ctd_dprintk_crit( > + "target %llx already in disconnect state\n", > + (emc_ctd_uint64_t)ctd_target->ctd_detect_name_bytes[0]); > + } > + error = 0; > + break; > + } > + } > + > + if (error) > + ctd_dprintk_crit("Error\n"); > + > + return error; > +} > + > +static int > +ctd_handle_target_addition(struct emc_ctd_v010_detect *io_detect, > + struct ctd_pci_private *ctd_private) > +{ > + int i; > + int error; > + struct ctd_target_info *ctd_target; > + struct ctd_host_info *ctd_host; > + > + error = -ENOMEM; > + ctd_host = ctd_private->host_private; > + > + ctd_dprintk("header addr -> %x key -> %llx\n", > + io_detect->emc_ctd_detect_header_address, > + io_detect->emc_ctd_v010_detect_key); > + > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > + ctd_target = &ctd_host->target[i]; > + > + /* detect header address is used to uniquely identify the target > + * for which the connect event has been posted by the server > + * check if this particular target is already recorded with the > + * client check if the recorded target is in the correct state > + * and if not found record this particular target in the list of > + * the detected targets > + */ > + > + if (ctd_target->ctd_detect.emc_ctd_v010_detect_key == > + io_detect->emc_ctd_v010_detect_key) { > + error = 0; > + ctd_dprintk("\n"); > + > + scsi_target_unblock(&ctd_target->starget->dev, > + SDEV_RUNNING); > + ctd_target->ctd_detect.emc_ctd_v010_detect_flags = > + io_detect->emc_ctd_v010_detect_flags; > + break; > + } > + > + /* End of list for the recorded targets in the client, so the > + * reported target is a new target being reported by the server > + * thus needs to be added to the list > + */ > + if (ctd_target->ctd_detect.emc_ctd_v010_detect_flags == 0x0) { > + error = 0; > + ctd_dprintk("\n"); > + ctd_target->ctd_detect = *io_detect; > + break; > + } > + } > + if (error) > + ctd_dprintk_crit("Error\n"); > + > + return error; > +} > + > +static int > +ctd_handle_source_addition(struct emc_ctd_v010_detect *io_detect, > + struct ctd_pci_private *ctd_private) > +{ > + ctd_dprintk("functionality not supported\n"); > + return -ENODEV; > +} > + > +static int > +ctd_handle_detect(struct emc_ctd_v010_detect *io_detect, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + > + /* > + * We post the detect event in the event queue and return. While > + * ctd_event_thread actually consumes the requests in the event queue. > + * This is done to serialize the consecutive detect requests (disconnect > + * followed by connect). This mechanism handles the situation when > + * multiple detect comes in a quick succession. Also, there is a > + * seperate thread and its queue for each adapter, so detect requests > + * for different adapters shall be handled seperately. > + */ > + error = ctd_post_event((union emc_ctd_v010_message *)io_detect, > + ctd_private); > + > + if (!error) > + atomic_long_inc(&ctd_private->hw_stats.what_in); > + > + return error; > + > +} > + > +static void > +ctd_handle_scsi_command(struct emc_ctd_v010_scsi_command *io_command, > + struct ctd_pci_private *ctd_private) > +{ > + ctd_dprintk_crit("unsupported\n"); > +} > + See below. > +static void > +ctd_handle_scsi_phase(struct emc_ctd_v010_scsi_phase *io_phase, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + unsigned long flags; > + struct scsi_cmnd *cmnd; > + struct ctd_request_private *request_private; > + > + /* check the phase flag to confirm we have received correct phase msg */ > + if (io_phase->emc_ctd_v010_scsi_phase_flags & > + EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET) { > + ctd_dprintk_crit("SCSI_PHASE_FLAG_TARGET not supported\n"); > + return; > + } > + > + if (!(io_phase->emc_ctd_v010_scsi_phase_flags & > + EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT)) { > + ctd_dprintk_crit("scsi_phase_flags %x invalid\n", > + io_phase->emc_ctd_v010_scsi_phase_flags); > + return; > + } > + > + request_private = (struct ctd_request_private *) > + ((uintptr_t)(io_phase->emc_ctd_v010_scsi_phase_opaque_rx)); > + > + error = ctd_scsi_response_sanity_check(request_private, ctd_private); > + > + if (error) > + return; > + > + cmnd = request_private->cmnd; > + > + ctd_dprintk_crit("SCSI_PHASE_FLAG_ABORT cmnd-> %p request -> %p\n", > + cmnd, request_private); > + > + atomic_long_inc(&ctd_private->hw_stats.abort_received); > + > + switch (request_private->io_state) { > + > + case CTD_IO_REQUEST_QUEUED: > + case CTD_IO_REQUEST_REQUEUED: > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + list_del(&request_private->list); > + atomic_long_dec(&ctd_private->hw_stats.active_io_count); > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + break; > + > + case CTD_IO_REQUEST_ABORTED: > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + list_del(&request_private->list); > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + > + case CTD_IO_REQUEST_REPLY_AWAITED: > + scsi_free_ctd_request_private(request_private, ctd_private); > + return; > + > + case CTD_IO_REQUEST_FREE: > + case CTD_IO_REQUEST_INVALID: > + default: > + ctd_dprintk_crit("opaque @ %p in unknown state %x\n", > + request_private, request_private->io_state); > + return; > + } > + > + cmnd->host_scribble = NULL; > + request_private->cmnd = NULL; > + request_private->io_state = CTD_IO_REQUEST_COMPLETED; > + scsi_free_ctd_request_private(request_private, ctd_private); > + > + /* error propagation to the SCSI midlayer */ > + scsi_translate_sam_code(cmnd, SAM_STAT_TASK_ABORTED); > + scsi_set_resid(cmnd, scsi_bufflen(cmnd)); > + cmnd->scsi_done(cmnd); > +} > + > +static void > +ctd_handle_response(union emc_ctd_v010_message *io_message, > + struct ctd_pci_private *ctd_private) > +{ > + struct emc_ctd_v010_header *msg_header; > + > + msg_header = &io_message->emc_ctd_v010_message_header; > + > + switch (msg_header->emc_ctd_v010_header_what) { > + > + case EMC_CTD_V010_WHAT_DETECT: > + ctd_handle_detect((struct emc_ctd_v010_detect *)io_message, > + ctd_private); > + break; > + case EMC_CTD_V010_WHAT_SCSI_COMMAND: > + ctd_handle_scsi_command( > + (struct emc_ctd_v010_scsi_command *)io_message, > + ctd_private); Print the unsopported message right here if you want to print it at all. > + break; > + case EMC_CTD_V010_WHAT_SCSI_PHASE: > + ctd_handle_scsi_phase( > + (struct emc_ctd_v010_scsi_phase *) io_message, > + ctd_private); > + break; > + case EMC_CTD_V010_WHAT_SCSI_RESPONSE: > + ctd_handle_scsi_response( > + (struct emc_ctd_v010_scsi_response *)io_message, > + ctd_private); > + break; > + default: > + ctd_dprintk_crit("unknown what -> %x ctd_private -> %p", > + msg_header->emc_ctd_v010_header_what, ctd_private); > + } > +} > + > + > +static int > +ctd_scsi_response_sanity_check(struct ctd_request_private *request_private, > + struct ctd_pci_private *ctd_private) > +{ > + int rc; > + unsigned long flags; > + struct scsi_cmnd *cmnd; > + struct ctd_request_private *request, *request_next; > + int outstanding_io_count = 0; > + > + rc = -EFAULT; > + > + /* check if the opaque is within the valid range */ > + if (!((request_private >= ctd_private->io_map) && > + (request_private < ctd_private->io_map_end))) { > + ctd_dprintk_crit( > + "ERROR request_private -> %p in invalid range\n", > + request_private); > + goto ctd_scsi_response_sanity_check_complete; > + } > + > + if (request_private) { > + cmnd = request_private->cmnd; > + if (cmnd) { > + /* check if the back pointer is valid before we declare > + * request sane > + */ > + if ((request_private == (struct ctd_request_private > + *)cmnd->host_scribble)) { > + rc = 0; > + goto ctd_scsi_response_sanity_check_complete; > + } > + } > + /* check if the request has already been completed to the SCSI > + * Midlayer > + */ > + else if ((request_private->io_state == > + CTD_IO_REQUEST_ABORTED) || > + (request_private->io_state == > + CTD_IO_REQUEST_REPLY_AWAITED)) { > + rc = 0; > + goto ctd_scsi_response_sanity_check_complete; > + } > + > + > + ctd_dprintk_crit( > + "ERROR cmnd -> %p mismatched request_private -> %p host_scribble -> %p requests send -> %ld responses received -> %ld state -> %s\n", > + cmnd, request_private, > + (cmnd ? cmnd->host_scribble : 0x0), > + ctd_private->hw_stats.requests_sent.counter, > + ctd_private->hw_stats.responses_received.counter, > + (request_private->io_state == > + CTD_IO_REQUEST_QUEUED ? > + "CTD_IO_REQUEST_QUEUED" : > + (request_private->io_state == > + CTD_IO_REQUEST_REQUEUED ? > + "CTD_IO_REQUEST_REQUEUED" : > + (request_private->io_state == > + CTD_IO_REQUEST_ABORTED ? > + "CTD_IO_REQUEST_ABORTED" : > + (request_private->io_state == > + CTD_IO_REQUEST_FREE ? > + "CTD_IO_REQUEST_FREE" : > + (request_private->io_state == > + CTD_IO_REQUEST_INVALID ? > + "CTD_IO_REQUEST_INVALID" : > + (request_private->io_state == > + CTD_IO_REQUEST_COMPLETED ? > + "CTD_IO_REQUEST_COMPLETED" : > + (request_private->io_state == > + CTD_IO_REQUEST_REPLY_AWAITED ? > + "CTD_IO_REQUEST_REPLY_AWAITED" : > + "UNKNOWN")))))))); > + > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + list_for_each_entry_safe(request, request_next, > + &ctd_private->aborted_io_list, list) { > + if (request == request_private) { > + ctd_dprintk_crit( > + "request_private -> %p in aborted_io_list\n", > + request_private); > + break; > + } > + } > + > + list_for_each_entry_safe(request, request_next, > + &ctd_private->queued_io_list, list) { > + if (request == request_private) { > + ctd_dprintk_crit( > + "request_private -> %p in queued_io_list\n", > + request_private); > + } > + outstanding_io_count++; > + } > + ctd_dprintk_crit("outstanding_io_count = %d\n", > + outstanding_io_count); > + > + list_for_each_entry_safe(request, request_next, > + &ctd_private->requeued_io_list, list) { > + if (request == request_private) { > + ctd_dprintk_crit( > + "request_private -> %p in requeued_io_list\n", > + request_private); > + break; > + } > + } > + > + list_for_each_entry_safe(request, request_next, > + &ctd_private->io_pool, list) { > + if (request == request_private) { > + ctd_dprintk_crit( > + "request_private -> %p in free io_pool\n", > + request_private); > + break; > + } > + } > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + } else { > + ctd_dprintk_crit("ERROR request_private -> NULL\n"); > + } > + > +ctd_scsi_response_sanity_check_complete: > + return rc; > +} > + > +static void > +ctd_handle_scsi_response(struct emc_ctd_v010_scsi_response *io_response, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + unsigned long flags; > + struct scsi_cmnd *cmnd; > + struct ctd_request_private *request_private; > + unsigned long long io_stats; > + > + request_private = (struct ctd_request_private *) > + ((uintptr_t)(io_response->emc_ctd_v010_scsi_response_opaque)); > + > + error = ctd_scsi_response_sanity_check(request_private, ctd_private); > + > + if (error) > + return; > + > + io_stats = ctd_read_tsc(); > + > + io_stats -= request_private->io_start_time; > + > + ctd_private->hw_stats.io_stats[ctd_private->hw_stats.io_stats_index++] = > + io_stats; > + > + if (ctd_private->hw_stats.io_stats_index == CTD_MAX_IO_STATS) > + ctd_private->hw_stats.io_stats_index = 0; > + /* > + * check if the IO context is in the aborted IO list > + */ > + > + /* Increment the responses_received stats */ > + atomic_long_inc(&ctd_private->hw_stats.responses_received); > + > + switch (request_private->io_state) { > + > + /* > + * The state of the request is important > + * CTD_IO_REQUEST_QUEUED : cmnd is still alive and valid in kernel > + * CTD_IO_REQUEST_ABORTED : cmnd has already been handled before > + * the response from the device and > + * only request needs to be cleaned > + * up from the abort list > + * CTD_IO_REQUEST_FREE : represents a state which is unhandled (unknown) > + * CTD_IO_REQUEST_REPLY_AWAITED : represents a state where abort > + * could not be sent by the timeout handler > + * CTD_IO_REQUEST_INVALID : represents a state which is > + * unhandled (unknown) > + */ > + > + case CTD_IO_REQUEST_QUEUED: > + case CTD_IO_REQUEST_REQUEUED: > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + list_del(&request_private->list); > + request_private->io_state = CTD_IO_REQUEST_COMPLETED; > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + break; > + > + case CTD_IO_REQUEST_ABORTED: > + /* > + * cmnd is already disassociated from the request private and IO > + * completed to the SCSI Midlayer by the timeout|abort handler > + * delink the request private from the aborted list and cleanup > + */ > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + list_del(&request_private->list); > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + > + case CTD_IO_REQUEST_REPLY_AWAITED: > + /* > + * return the context back to the io_pool for its reuse > + */ > + request_private->io_state = CTD_IO_REQUEST_COMPLETED; > + scsi_free_ctd_request_private(request_private, ctd_private); > + /* IO already completed to the Midlayer */ > + return; > + > + case CTD_IO_REQUEST_FREE: > + case CTD_IO_REQUEST_INVALID: > + default: > + ctd_dprintk_crit( > + "opaque @ %p in unknown state %x\n", > + request_private, request_private->io_state); > + return; > + > + } > + > + /* Decrement active_io_count only when request is still queued. */ > + atomic_long_dec(&ctd_private->hw_stats.active_io_count); > + > + cmnd = request_private->cmnd; > + > + cmnd->result = 0; > + > + scsi_translate_sam_code(cmnd, > + io_response->emc_ctd_v010_scsi_response_status); > + > + scsi_set_resid(cmnd, scsi_bufflen(cmnd) - > + io_response->emc_ctd_v010_scsi_response_data_size); > + > + if (io_response->emc_ctd_v010_scsi_response_flags & > + EMC_CTD_V010_SCSI_RESPONSE_FLAG_FAILED) { > + ctd_dprintk_crit( > + "cmnd = %p CTDCM_FAILED channel:%d id:%d lun:%lld] status = %x\n", > + cmnd, cmnd->device->channel, cmnd->device->id, > + cmnd->device->lun, > + io_response->emc_ctd_v010_scsi_response_status); > + > + set_host_byte(cmnd, DID_ERROR); > + if (ctd_debug) { > + scsi_print_command(cmnd); > + scsi_print_result(cmnd, NULL, FAILED); > + scsi_print_sense(cmnd); > + } > + } > + if (io_response->emc_ctd_v010_scsi_response_flags & > + EMC_CTD_V010_SCSI_RESPONSE_FLAG_SENSE) { > + unsigned char *sense_data; > + unsigned char sense_data_length; > + > + sense_data = io_response->emc_ctd_scsi_response_extra; > + sense_data_length = > + io_response->emc_ctd_v010_scsi_response_extra_size; > + > + memcpy(cmnd->sense_buffer, sense_data, sense_data_length); > + > + set_driver_byte(cmnd, DRIVER_SENSE); > + if (ctd_debug) { > + scsi_print_command(cmnd); > + scsi_print_result(cmnd, "emcctd sense", SUCCESS); > + scsi_print_sense(cmnd); > + } > + } > + > + if ((io_response->emc_ctd_v010_scsi_response_status == > + SAM_STAT_TASK_SET_FULL) || > + (io_response->emc_ctd_v010_scsi_response_status == > + SAM_STAT_BUSY)) { > + > + ctd_dprintk( > + "QUEUE DEPTH change for channel:%d id:%d lun:%lld] active io count = %lx\n", > + cmnd->device->channel, cmnd->device->id, > + cmnd->device->lun, > + ctd_private->hw_stats.active_io_count.counter); > + > + scsi_track_queue_full(cmnd->device, > + ctd_private->hw_stats.active_io_count.counter); > + } > + > + > + cmnd->host_scribble = NULL; > + > + scsi_free_ctd_request_private(request_private, ctd_private); > + > + cmnd->scsi_done(cmnd); > +} > + > +static void > +ctd_scsi_transfer_info(unsigned char *cmd, unsigned long long *lba, > + unsigned int *num, unsigned int *ei_lba) > +{ > + *ei_lba = 0; > + > + switch (*cmd) { > + > + case VARIABLE_LENGTH_CMD: > + *lba = (u64)cmd[19] | (u64)cmd[18] << 8 | > + (u64)cmd[17] << 16 | (u64)cmd[16] << 24 | > + (u64)cmd[15] << 32 | (u64)cmd[14] << 40 | > + (u64)cmd[13] << 48 | (u64)cmd[12] << 56; > + > + *ei_lba = (u32)cmd[23] | (u32)cmd[22] << 8 | > + (u32)cmd[21] << 16 | (u32)cmd[20] << 24; > + > + *num = (u32)cmd[31] | (u32)cmd[30] << 8 | (u32)cmd[29] << 16 | > + (u32)cmd[28] << 24; > + break; > + > + case WRITE_SAME_16: > + case WRITE_16: > + case READ_16: > + *lba = (u64)cmd[9] | (u64)cmd[8] << 8 | > + (u64)cmd[7] << 16 | (u64)cmd[6] << 24 | > + (u64)cmd[5] << 32 | (u64)cmd[4] << 40 | > + (u64)cmd[3] << 48 | (u64)cmd[2] << 56; > + > + *num = (u32)cmd[13] | (u32)cmd[12] << 8 | (u32)cmd[11] << 16 | > + (u32)cmd[10] << 24; > + break; > + case WRITE_12: > + case READ_12: > + *lba = (u32)cmd[5] | (u32)cmd[4] << 8 | (u32)cmd[3] << 16 | > + (u32)cmd[2] << 24; > + > + *num = (u32)cmd[9] | (u32)cmd[8] << 8 | (u32)cmd[7] << 16 | > + (u32)cmd[6] << 24; > + break; > + case WRITE_SAME: > + case WRITE_10: > + case READ_10: > + case XDWRITEREAD_10: > + *lba = (u32)cmd[5] | (u32)cmd[4] << 8 | (u32)cmd[3] << 16 | > + (u32)cmd[2] << 24; > + > + *num = (u32)cmd[8] | (u32)cmd[7] << 8; > + break; > + case WRITE_6: > + case READ_6: > + *lba = (u32)cmd[3] | (u32)cmd[2] << 8 | > + (u32)(cmd[1] & 0x1f) << 16; > + *num = (cmd[4] == 0) ? 256 : cmd[4]; > + break; > + default: > + break; > + } > +} > + > + > +static int > +ctd_initiator_validate_sgl(struct scsi_cmnd *cmnd, > + struct emc_ctd_v010_scsi_command *ctd_request, > + struct ctd_pci_private *ctd_private) > +{ > + int i; > + int error; > + struct emc_ctd_v010_sgl *sgl_current; > + struct emc_ctd_v010_sgl *sgl_extended; > + emc_ctd_uint64_t sgl_count; > + unsigned int sgl_buffersize; > + unsigned long long lba; > + unsigned int num; > + unsigned int ei_lba; > + > + lba = 0; > + num = 0; > + ei_lba = 0; > + error = -EINVAL; > + sgl_count = 0; > + sgl_extended = &ctd_request->emc_ctd_scsi_command_sgl[0]; > + > + if (ctd_request->emc_ctd_v010_scsi_command_flags & > + EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL) { > + emc_ctd_uint64_t extented_sgl_size; > + > + sgl_current = phys_to_virt( > + ((emc_ctd_uint64_t)sgl_extended->emc_ctd_v010_sgl_paddr_32_63 << 32) | > + ((emc_ctd_uint64_t)sgl_extended->emc_ctd_v010_sgl_paddr_0_31 & 0xFFFFFFFF)); > + > + extented_sgl_size = sgl_extended->emc_ctd_v010_sgl_size; > + > + do_div(extented_sgl_size, sizeof(struct emc_ctd_v010_sgl)); > + > + sgl_count = extented_sgl_size; > + } else { > + sgl_current = sgl_extended; > + while (sgl_current < sgl_extended + > + EMC_CTD_V010_SGL_IMMEDIATE_MAX) { > + if (sgl_current->emc_ctd_v010_sgl_size == 0) > + break; > + > + sgl_current++; > + sgl_count++; > + } > + sgl_current = sgl_extended; > + } > + > + if (scsi_sg_count(cmnd) != sgl_count) { > + scsi_print_command(cmnd); > + ctd_dprintk_crit( > + "Mismatched cmnd_sgl_count %d sgl_count = %lld sgl_size = %d\n", > + scsi_sg_count(cmnd), > + sgl_count, sgl_extended->emc_ctd_v010_sgl_size); > + } > + > + if (sgl_count) { > + struct scatterlist *cmnd_sg; > + int cmnd_sg_count = scsi_sg_count(cmnd); > + > + if (cmnd_sg_count > EMC_CTD_V010_SGL_IMMEDIATE_MAX) { > + if ((ctd_request->emc_ctd_v010_scsi_command_flags & > + EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL) == 0) { > + scsi_print_command(cmnd); > + ctd_dprintk_crit( > + "scsi_sg_count = %d but EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL not set\n", > + scsi_sg_count(cmnd)); > + > + cmnd_sg_count = EMC_CTD_V010_SGL_IMMEDIATE_MAX; > + } > + } > + > + scsi_for_each_sg(cmnd, cmnd_sg, cmnd_sg_count, i) { > + int cmnd_sg_size; > + emc_ctd_uint64_t cmnd_buffer_pfn; > + struct page *cmnd_page; > + emc_ctd_uint32_t cmnd_offset; > + emc_ctd_uint32_t sgl_size; > + emc_ctd_uint64_t buffer_pfn; > + > + cmnd_page = sg_page(cmnd_sg); > + cmnd_buffer_pfn = page_to_phys(cmnd_page); > + cmnd_sg_size = cmnd_sg->length; > + cmnd_offset = cmnd_sg->offset; > + > + > + sgl_size = (sgl_current + i)->emc_ctd_v010_sgl_size; > + buffer_pfn = (((emc_ctd_uint64_t) > + ((sgl_current + i)->emc_ctd_v010_sgl_paddr_32_63) << 32) | > + ((emc_ctd_uint64_t)((sgl_current + i)->emc_ctd_v010_sgl_paddr_0_31) & 0xFFFFFFFF)); > + > + if ((sgl_size != cmnd_sg_size) || > + (buffer_pfn != (cmnd_buffer_pfn + > + cmnd_offset))) { > + scsi_print_command(cmnd); > + ctd_dprintk_crit( > + "Mismatch @ i = %d cmnd_sg_size = %d cmnd_buffer_pfn = %llx sgl_size = %d buffer_pfn = %llx\n", > + i, cmnd_sg_size, cmnd_buffer_pfn, > + sgl_size, buffer_pfn); > + } > + } > + goto ctd_initiator_validate_sgl_end; > + } > + > + for (i = 0, sgl_buffersize = 0; i < sgl_count; i++, sgl_current++) > + sgl_buffersize += sgl_current->emc_ctd_v010_sgl_size; > + > + if (scsi_bufflen(cmnd) && (sgl_buffersize != scsi_bufflen(cmnd))) { > + scsi_print_command(cmnd); > + ctd_dprintk_crit("Mismatched buffer size %d %d\n", > + scsi_bufflen(cmnd), sgl_buffersize); > + goto ctd_initiator_validate_sgl_end; > + } > + > + ctd_scsi_transfer_info((unsigned char *)cmnd->cmnd, &lba, > + &num, &ei_lba); > + > + if (num && (sgl_buffersize != (num * 512))) { > + scsi_print_command(cmnd); > + ctd_dprintk_crit("Mismatched buffer size %d %d\n", > + (num * 512), sgl_buffersize); > + goto ctd_initiator_validate_sgl_end; > + } > + error = 0; > +ctd_initiator_validate_sgl_end: > + return error; > +} > + > +static int > +ctd_initiator_translate_sgl(struct scsi_cmnd *cmnd, > + struct emc_ctd_v010_scsi_command *ctd_request, > + struct ctd_pci_private *ctd_private) > +{ > + int i; > + int size; > + int error; > + int sg_count; > + int rq_count; > + int embedded_sgl_count; > + int embedded_sgl_index; > + struct ctd_request_private *request_private; > + emc_ctd_uint64_t buffer_pfn; > + > + error = FAILED; > + i = size = 0; > + embedded_sgl_index = 0; > + embedded_sgl_count = EMC_CTD_V010_SGL_IMMEDIATE_MAX; > + rq_count = 0; > + > + request_private = (struct ctd_request_private *)cmnd->host_scribble; > + sg_count = scsi_sg_count(cmnd); > + > + if (sg_count > embedded_sgl_count) { > + struct scatterlist *sg; > + struct emc_ctd_v010_sgl *sgl_current; > + struct emc_ctd_v010_sgl *sgl_extended; > + > + request_private->sgllist_page_order = > + get_order((sizeof(struct emc_ctd_v010_sgl) * sg_count)); > + request_private->sgllist_page = > + alloc_pages(GFP_ATOMIC | __GFP_COMP | __GFP_NOWARN, > + request_private->sgllist_page_order); > + > + if (!request_private->sgllist_page) { > + ctd_dprintk_crit("alloc_page failure\n"); > + goto scsi_initiator_translate_sgl_end; > + } > + > + sgl_current = (struct emc_ctd_v010_sgl *) > + page_address(request_private->sgllist_page); > + > + scsi_for_each_sg(cmnd, sg, sg_count, i) { > +#if defined(__VMKLNX__) > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > + sg->cursgel->addr & 0xFFFFFFFF; > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 = > + sg->cursgel->addr >> 32; > + sgl_current->emc_ctd_v010_sgl_size = sg_dma_len(sg); > +#else __VMKLNX__ won't be defined here. > + struct page *page; > + > + page = sg_page(sg); > + buffer_pfn = page_to_phys(page); > + size += sg->length; > + > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > + (buffer_pfn + sg->offset) & 0xFFFFFFFF; > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 = > + (buffer_pfn + sg->offset) >> 32; > + sgl_current->emc_ctd_v010_sgl_size = sg->length; > +#endif > + sgl_current++; > + rq_count++; > + } > + sgl_extended = &ctd_request->emc_ctd_scsi_command_sgl[0]; > + buffer_pfn = page_to_phys(request_private->sgllist_page); > + > + sgl_extended->emc_ctd_v010_sgl_paddr_0_31 = > + buffer_pfn & 0xFFFFFFFF; > + sgl_extended->emc_ctd_v010_sgl_paddr_32_63 = buffer_pfn >> 32; > + sgl_extended->emc_ctd_v010_sgl_size = > + rq_count * sizeof(struct emc_ctd_v010_sgl); > + ctd_request->emc_ctd_v010_scsi_command_flags |= > + EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL; > + } else { > + struct scatterlist *sg; > + struct emc_ctd_v010_sgl *sgl_current; > + > + sgl_current = &ctd_request->emc_ctd_scsi_command_sgl[0]; > + > + scsi_for_each_sg(cmnd, sg, sg_count, i) { Same here. > +#if defined(__VMKLNX__) > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > + sg->cursgel->addr & 0xFFFFFFFF; > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 = > + sg->cursgel->addr >> 32; > + sgl_current->emc_ctd_v010_sgl_size = sg_dma_len(sg); > +#else > + struct page *page; > + > + size += sg->length; > + page = sg_page(sg); > + buffer_pfn = page_to_phys(page); > + > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > + (buffer_pfn + sg->offset) & 0xFFFFFFFF; > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 = > + (buffer_pfn + sg->offset) >> 32; > + sgl_current->emc_ctd_v010_sgl_size = sg->length; > +#endif > + > + sgl_current++; > + rq_count++; > + } > + } > + error = SUCCESS; > + > + if (ctd_debug) > + ctd_initiator_validate_sgl(cmnd, ctd_request, ctd_private); > + > +scsi_initiator_translate_sgl_end: > + return error; > +} > + > + > +void > +ctd_initiator_translate_lun(struct scsi_cmnd *cmnd, > + struct emc_ctd_v010_scsi_command *ctd_request) > +{ > + > + int i; > + union ctd_scsi_lun { > + emc_ctd_uint64_t scsi_lun_64; > + emc_ctd_uint8_t scsi_lun_8[8]; > + } ctd_scsi_lun_conversion; > + > + ctd_scsi_lun_conversion.scsi_lun_64 = cpu_to_be64(cmnd->device->lun); > + for (i = 0; i < sizeof(ctd_scsi_lun_conversion); i++) { > + ctd_request->emc_ctd_v010_scsi_command_lun[i] = > + ctd_scsi_lun_conversion.scsi_lun_8[i]; > + } > +} > + > +static int > +ctd_initiator_translate_request(struct scsi_cmnd *cmnd, > + struct emc_ctd_v010_scsi_command *ctd_request, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + struct ctd_dev_info *ctd_device; > + struct ctd_request_private *request_private; > + int scsi_cdb_size; > + > + error = FAILED; > + request_private = NULL; > + ctd_device = cmnd->device->hostdata; > + > + if (!(ctd_device->ctd_target_detect->emc_ctd_v010_detect_flags & > + EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) { > + goto scsi_initiator_translate_request_end; > + } > + memset(ctd_request, 0x0, sizeof(struct emc_ctd_v010_scsi_command)); > + > + ctd_request->emc_ctd_scsi_command_header_address = > + ctd_device->ctd_target_detect->emc_ctd_detect_header_address; > + ctd_request->emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_minor = > + EMCCTD_V010_PROTOCOL_MINOR_VERSION; > + ctd_request->emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_what = > + EMC_CTD_V010_WHAT_SCSI_COMMAND; > + ctd_request->emc_ctd_v010_scsi_command_flags |= > + ((cmnd->sc_data_direction == DMA_FROM_DEVICE) ? > + EMC_CTD_V010_SCSI_COMMAND_FLAG_SOURCE : > + ((cmnd->sc_data_direction == > + DMA_TO_DEVICE) ? 0 : 0)); > + > + request_private = ctd_acquire_request(ctd_private); > + > + if (request_private == NULL) { > + ctd_dprintk_crit("ctd_acquire_request failure\n"); > + goto scsi_initiator_translate_request_end; > + } > + > + request_private->cmnd = cmnd; > + > + cmnd->host_scribble = (unsigned char *)request_private; > + > + ctd_request->emc_ctd_v010_scsi_command_data_size = scsi_bufflen(cmnd); > + > + ctd_request->emc_ctd_v010_scsi_command_opaque = > + (uintptr_t)request_private; > + > + ctd_initiator_translate_lun(cmnd, ctd_request); > + > + scsi_cdb_size = cmnd->cmd_len; > + > + if (scsi_cdb_size <= > + sizeof(ctd_request->emc_ctd_v010_scsi_command_cdb)) { > + memcpy(ctd_request->emc_ctd_v010_scsi_command_cdb, > + cmnd->cmnd, scsi_cdb_size); > + } else { > + ctd_dprintk_crit("unsupported scsi cdb size = %d\n", > + scsi_cdb_size); > + goto scsi_initiator_translate_request_end; > + } > + > + error = ctd_initiator_translate_sgl(cmnd, ctd_request, ctd_private); > + > +scsi_initiator_translate_request_end: > + if (error == FAILED) { > + cmnd->host_scribble = NULL; > + if (request_private) { > + scsi_free_ctd_request_private(request_private, > + ctd_private); > + } > + } > + return error; > +} > + > + > +static int > +ctd_hw_execute_command(struct scsi_cmnd *cmnd, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + unsigned long flags; > + struct emc_ctd_v010_scsi_command ctd_request; > + > + error = ctd_initiator_translate_request(cmnd, &ctd_request, > + ctd_private); > + > + if (error == SUCCESS) { > + struct ctd_request_private *request_private; > + > + request_private = > + (struct ctd_request_private *)cmnd->host_scribble; > + > + /* lock ordering ... io_mgmt_lock followed by isr_lock > + * ensures that request placed in legitimate queue ... > + * so that response finds in the correct queue > + */ > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + request_private->io_start_time = ctd_read_tsc(); > + > + error = ctd_hw_enqueue_request( > + (union emc_ctd_v010_message *)&ctd_request, > + ctd_private); > + if (error == SUCCESS) { > + list_add_tail(&request_private->list, > + &ctd_private->queued_io_list); > + atomic_long_inc(&ctd_private->hw_stats.requests_sent); > + atomic_long_inc(&ctd_private->hw_stats.active_io_count); > + } > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + > + if (error != SUCCESS) { > + ctd_dprintk_crit("ctd_hw_enqueue_request error\n"); > + scsi_free_ctd_request_private( > + (struct ctd_request_private *) > + (uintptr_t) > + (ctd_request.emc_ctd_v010_scsi_command_opaque), > + ctd_private); > + } > + } > + > + return error; > +} > + > +static int > +ctd_hw_enqueue_request(union emc_ctd_v010_message *ctd_request, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + unsigned long flags; > + union emc_ctd_v010_message *ctd_request_block; > + > + spin_lock_irqsave(&ctd_private->isr_lock, flags); > + > + /* > + * check if any space is available in the array > + */ > + ctd_request_block = > + ((((ctd_private->request_producer_index + 1) % > + ctd_private->pci_request_queue_size) == > + ctd_private->request_consumer_index) ? > + NULL : > + (ctd_private->pci_request_array + > + ctd_private->request_producer_index)); > + > + > + error = (ctd_request_block ? SUCCESS : FAILED); > + > + if (error == SUCCESS) { > + *ctd_request_block = *ctd_request; > + > + ctd_private->request_producer_index = > + ((ctd_private->request_producer_index + 1) % > + ctd_private->pci_request_queue_size); > + > + } > + > + spin_unlock_irqrestore(&ctd_private->isr_lock, flags); > + > + return error; > +} > + > +static int > +ctd_hw_dequeue_response(union emc_ctd_v010_message *ctd_response, > + struct ctd_pci_private *ctd_private) > +{ > + int rc; > + unsigned long flags; > + union emc_ctd_v010_message *ctd_response_block; > + > + /* protect ourselves from another instance */ > + spin_lock_irqsave(&ctd_private->isr_lock, flags); > + > + ctd_response_block = > + ((ctd_private->response_consumer_index == > + ctd_private->response_producer_index) ? NULL : > + (ctd_private->pci_response_array + > + ctd_private->response_consumer_index)); > + > + rc = ctd_response_block ? SUCCESS : FAILED; > + > + if (rc == SUCCESS) { > + *ctd_response = *ctd_response_block; > + ctd_private->response_consumer_index = > + ((ctd_private->response_consumer_index + 1) % > + ctd_private->pci_response_queue_size); > + } else { > + ctd_check_error_condition(ctd_private->pci_dev); > + } > + > + spin_unlock_irqrestore(&ctd_private->isr_lock, flags); > + > + return rc; > + > +} > + > +static int > +ctd_xmit_command(struct scsi_cmnd *cmnd, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + > + cmnd->result = DID_OK; > + error = ctd_hw_execute_command(cmnd, ctd_private); > + return error; > +} > + > + > +static int > +ctd_queuecommand_lck(struct scsi_cmnd *cmnd, void (*done)(struct scsi_cmnd *)) > +{ > + int error; > + struct ctd_host_info *ctd_host; > + struct ctd_pci_private *ctd_private; > + > + error = 0; > + ctd_host = shost_priv(cmnd->device->host); > + ctd_private = pci_get_drvdata(ctd_host->pci_dev); > + > + switch (ctd_private->hw_state) { > + > + case CTD_HW_STATE_INITIALIZED: > + cmnd->scsi_done = done; > + if (ctd_xmit_command(cmnd, ctd_private) == SUCCESS) > + break; > + > + case CTD_HW_STATE_DISABLED: > + cmnd->scsi_done = done; > + scsi_translate_sam_code(cmnd, SAM_STAT_TASK_ABORTED); > + scsi_set_resid(cmnd, scsi_bufflen(cmnd)); > + cmnd->scsi_done(cmnd); > + break; > + default: > + error = SCSI_MLQUEUE_HOST_BUSY; > + } > + > + return error; > +} > + > +static int > +ctd_abort_handler(struct scsi_cmnd *cmnd) > +{ > + ctd_dprintk_crit("SCSI cmnd -> %p ERROR\n", > + cmnd); > + return SUCCESS; > +} > + > + > +static int > +ctd_target_alloc(struct scsi_target *starget) > +{ > + int error; > + struct ctd_target_info *ctd_target; > + struct ctd_host_info *ctd_host; > + > + error = -ENODEV; > + ctd_host = shost_priv(dev_to_shost(&starget->dev)); > + > + ctd_dprintk("starget -> %p id -> %x\n", > + starget, starget->id); > + > + ctd_target = &ctd_host->target[starget->id]; > + > + /* check for the connection status in the detect flag > + * also check if the target already registered with the SCSI midlayer > + */ > + if (ctd_target->starget == NULL && > + (ctd_target->ctd_detect.emc_ctd_v010_detect_flags & > + EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) { > + error = 0; > + ctd_target->starget = starget; > + starget->hostdata = ctd_target; > + } else { > + if (ctd_target->starget != starget) { > + ctd_dprintk_crit( > + "failure ctd_target->starget %p and starget %p dissimilar\n", > + ctd_target->starget, starget); > + } else { > + ctd_dprintk_crit("failure starget %p unexpected\n", > + starget); > + } > + } > + > + return error; > +} > + > +void > +ctd_target_destroy(struct scsi_target *starget) > +{ > + int i; > + int error; > + struct ctd_target_info *ctd_target; > + struct ctd_host_info *ctd_host; > + > + error = -ENODEV; > + ctd_host = shost_priv(dev_to_shost(&starget->dev)); > + > + ctd_dprintk_crit("starget @ id = %x\n", starget->id); > + > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > + ctd_target = &ctd_host->target[i]; > + if (ctd_target->starget == starget) { > + ctd_target->starget = NULL; > + error = 0; > + break; > + } > + } > + > + if (error) { > + ctd_dprintk_crit("failure for starget @ id = %x\n", > + starget->id); > + } No need for the parentheses here. > +} > + > +static int > +ctd_slave_configure(struct scsi_device *sdevice) > +{ > + int error; > + struct ctd_dev_info *ctd_device; > + struct ctd_host_info *ctd_host; > + > + error = 0; > + ctd_host = shost_priv(sdevice->host); > + ctd_device = sdevice->hostdata; > + > + /* tune the block layer to generate timout > + * for the requests queued and reply is awaited > + */ > + blk_queue_rq_timeout(sdevice->request_queue, EMCCTD_REQUEST_TIMEOUT); > + > + return error; > +} > + > + > + > +static int > +ctd_slave_alloc(struct scsi_device *sdev) > +{ > + int error; > + struct ctd_host_info *ctd_host; > + struct ctd_target_info *ctd_target; > + struct ctd_dev_info *ctd_device; > + > + error = -ENOMEM; > + ctd_host = shost_priv(sdev->host); > + sdev->host->max_cmd_len = EMCCTD_V010_MAX_CDB_SIZE; > + > + ctd_device = (struct ctd_dev_info *) > + kzalloc(sizeof(struct ctd_dev_info), GFP_ATOMIC); > + if (ctd_device == NULL) > + goto ctd_slave_alloc_end; > + > + ctd_target = &ctd_host->target[sdev->id]; > + if (ctd_target->starget) > + ctd_device->ctd_target_detect = &ctd_target->ctd_detect; > + > + if (ctd_device->ctd_target_detect) { > + error = 0; > + ctd_device->ctd_host = ctd_host; > + ctd_device->ctd_target = ctd_target; > + sdev->hostdata = ctd_device; > + queue_flag_set_unlocked(QUEUE_FLAG_BIDI, sdev->request_queue); > + } else { > + kfree(ctd_device); > + error = -ENODEV; > + } > + > +ctd_slave_alloc_end: > + > + if (error) { > + ctd_dprintk_crit("channel = %x id= %x error = %x\n", > + sdev->channel, sdev->id, error); > + } Same goes here. > + return error; > +} > + > +static void > +ctd_slave_destroy(struct scsi_device *sdev) > +{ > + struct ctd_dev_info *dev_info; > + > + dev_info = sdev->hostdata; > + kfree(dev_info); > +} > + > +static enum blk_eh_timer_return > +ctd_timeout_handler(struct scsi_cmnd *cmd) > +{ > + int requeue_error; > + unsigned long flags; > + enum blk_eh_timer_return error; > + struct ctd_host_info *ctd_host; > + struct ctd_pci_private *ctd_private; > + struct ctd_request_private *request; > + unsigned long long tsc_val; > + > + requeue_error = FAILED; > + error = BLK_EH_NOT_HANDLED; > + ctd_host = shost_priv(cmd->device->host); > + ctd_private = pci_get_drvdata(ctd_host->pci_dev); > + > + > + request = (struct ctd_request_private *)cmd->host_scribble; > + > + tsc_val = ctd_read_tsc(); > + > + tsc_val -= request->io_start_time; > + > + ctd_dprintk_crit("cmnd -> %p request -> %p, tsc -> %lld\n", > + cmd, request, tsc_val); > + > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + if (request && request->io_timeout < EMCCTD_MAX_RETRY) { > + struct ctd_dev_info *ctd_device; > + > + switch (request->io_state) { > + > + /* check if the IO in the queued_io_list */ > + case CTD_IO_REQUEST_QUEUED: > + /* check if the IO is already requeued */ > + case CTD_IO_REQUEST_REQUEUED: > + /* remove the old IO context from the requeued_io_list > + * or queued_io_list > + */ > + list_del(&request->list); > + atomic_long_dec(&ctd_private->hw_stats.active_io_count); > + > + ctd_device = cmd->device->hostdata; > + > + if (!(ctd_device->ctd_target_detect->emc_ctd_v010_detect_flags & > + EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) { > + ctd_dprintk_crit("device diconnected\n"); > + } else { > + union emc_ctd_v010_message ctd_message; > + struct emc_ctd_v010_scsi_phase *ctd_phase; > + > + memset(&ctd_message, 0x0, > + sizeof(union emc_ctd_v010_message)); > + ctd_phase = > + (struct emc_ctd_v010_scsi_phase *) & > + ctd_message.emc_ctd_v010_message_scsi_phase; > + > + /* Need to acertain if this is how an IO is > + * aborted by the specification > + */ > + > + /* > + * OPT-438489 the phase flag needs to be > + * initialized with PHASE_FLAG_TARGET If EMC > + * CTD V010 SCSI PHASE FLAG TARGET is set, the > + * message receiver is the target, otherwise the > + * message receiver is the initiator. If EMC CTD > + * V010 SCSI PHASE FLAG ABORT is set, the SCSI > + * command is aborted. > + */ > + ctd_phase->emc_ctd_v010_scsi_phase_flags = > + EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT | > + EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET; > + > + ctd_phase->emc_ctd_v010_scsi_phase_opaque_tx = > + (uintptr_t)request; > + ctd_phase->emc_ctd_v010_scsi_phase_opaque_rx = > + (-1); > + ctd_phase->emc_ctd_scsi_phase_header_what = > + EMC_CTD_V010_WHAT_SCSI_PHASE; > + ctd_phase->emc_ctd_scsi_phase_header_minor = > + EMCCTD_V010_PROTOCOL_MINOR_VERSION; > + ctd_phase->emc_ctd_scsi_phase_header_address = > + ctd_device->ctd_target_detect->emc_ctd_detect_header_address; > + > + requeue_error = > + ctd_hw_enqueue_request(&ctd_message, > + ctd_private); > + } > + > + > + if (requeue_error != SUCCESS) { > + /* add the IO context to requeued_io_list */ > + /* Client would try to abort the request in next > + * timeout (after 20 seconds) > + */ > + request->io_state = CTD_IO_REQUEST_REQUEUED; > + list_add_tail(&request->list, > + &ctd_private->requeued_io_list); > + request->io_timeout++; > + error = BLK_EH_RESET_TIMER; > + } else { > + request->cmnd = NULL; > + cmd->host_scribble = NULL; > + > + request->io_state = CTD_IO_REQUEST_ABORTED; > + request->purge_lifetime = > + jiffies + EMCCTD_OPAQUE_PURGE_WAITTIME; > + list_add_tail(&request->list, > + &ctd_private->aborted_io_list); > + atomic_long_inc(&ctd_private->hw_stats.abort_sent); > + > + /* error propagation to the SCSI midlayer */ > + scsi_translate_sam_code(cmd, > + SAM_STAT_TASK_ABORTED); > + scsi_set_resid(cmd, scsi_bufflen(cmd)); > + > + /* indicate no more requeue of this particular > + * IO is needed > + */ > + error = BLK_EH_HANDLED; > + } > + break; > + > + default: > + ctd_dprintk_crit( > + "request @ %p in unexpected state %x\n", > + request, request->io_state); > + break; > + } > + } else if (request) { > + ctd_dprintk_crit("cmd %p timeout completed io_state %x\n", > + cmd, request->io_state); > + > + /* remove the old IO context from the requeued_io_list */ > + list_del(&request->list); > + > + /* > + * break the context between the cmnd and the request and > + * request on the requeue_io_list cannot be reused until > + * server replies for the same > + */ > + request->cmnd = NULL; > + cmd->host_scribble = NULL; > + > + /* error propagation to the SCSI midlayer */ > + scsi_translate_sam_code(cmd, SAM_STAT_TASK_ABORTED); > + scsi_set_resid(cmd, scsi_bufflen(cmd)); > + > + /* we can deallocate the context only once we receive > + * any reply from the server > + */ > + request->io_state = CTD_IO_REQUEST_REPLY_AWAITED; > + > + /* indicate no more requeue of this particular IO is needed */ > + error = BLK_EH_HANDLED; > + } else { > + ctd_dprintk_crit( > + "cmnd -> %p request -> NULL error !!!\n", cmd); > + } > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + > + return error; > +} > + > + > +static int > +ctd_ITnexus_handler(struct ctd_pci_private *ctd_private) > +{ > + int i; > + int error; > + struct ctd_target_info *ctd_target; > + struct ctd_host_info *ctd_host; > + > + error = 0; > + > + ctd_host = (struct ctd_host_info *)ctd_private->host_private; > + > + ctd_dprintk_crit("ctd_private -> %p\n", ctd_private); > + > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > + union emc_ctd_v010_message ctd_message; > + struct emc_ctd_v010_detect *ctd_detect; > + > + ctd_target = &ctd_host->target[i]; > + > + switch (ctd_target->detect_completed) { > + > + case EMCCTD_TARGET_DETECT_NOT_COMPLETED: > + if (!ctd_target->ctd_detect.emc_ctd_v010_detect_flags) > + break; > + > + /* The id defined by the SCSI Midlayer should match the > + * index as this routine is indirectly invoked by the > + * delayed work mechanism > + */ > + > + ctd_dprintk_crit("ctd_target -> %p index = %x\n", > + ctd_target, i); > + ctd_dprintk_crit("key -> %llx header -> %x\n", > + ctd_target->ctd_detect.emc_ctd_v010_detect_key, > + ctd_target->ctd_detect.emc_ctd_detect_header_address); > + > + memset(&ctd_message, 0x0, > + sizeof(union emc_ctd_v010_message)); > + > + ctd_detect = &ctd_message.emc_ctd_v010_message_detect; > + ctd_detect->emc_ctd_v010_detect_flags = 0x0; > + ctd_detect->emc_ctd_v010_detect_key = > + ctd_target->ctd_detect.emc_ctd_v010_detect_key; > + ctd_detect->emc_ctd_detect_header_what = > + EMC_CTD_V010_WHAT_DETECT; > + ctd_detect->emc_ctd_detect_header_minor = > + EMCCTD_V010_PROTOCOL_MINOR_VERSION; > + ctd_detect->emc_ctd_detect_header_address = > + ctd_target->ctd_detect.emc_ctd_detect_header_address; > + > + if (ctd_hw_enqueue_request(&ctd_message, > + ctd_private) == SUCCESS) { > + ctd_dprintk_crit("ctd_target -> %p\n", > + ctd_target); > + ctd_target->detect_completed = > + EMCCTD_TARGET_DETECT_COMPLETED; > + > + atomic_long_inc( > + &ctd_private->hw_stats.what_out); > + } else { > + ctd_dprintk_crit( > + "ctd_target -> %p ctd_hw_enqueue_request failure\n", > + ctd_target); > + error = -EAGAIN; > + break; > + } > + > + case EMCCTD_TARGET_DETECT_COMPLETED: > + /* Disconnect case ... we need to remove the associated > + * objects from SCSI midlayer > + */ > + if (!ctd_target->ctd_detect.emc_ctd_v010_detect_flags) { > + ctd_dprintk_crit("ctd_target -> %p\n", > + ctd_target); > + > + ctd_clear_io_queue(ctd_private); > + > + if (ctd_target->starget) { > + /* the following attempts to clean the SCSI > + * midlayer objects > + */ > + scsi_target_block( > + &ctd_target->starget->dev); > + scsi_target_unblock( > + &ctd_target->starget->dev, > + SDEV_TRANSPORT_OFFLINE); > + /* > + * Target object might still be active > + * in case its not reaped completely (as > + * with LVM) thus might be reused when > + * the link reconnects back (OPT 443532) > + */ > + scsi_remove_target( > + &ctd_target->starget->dev); > + } else { > + ctd_dprintk_crit( > + "starget already null\n"); > + } > + > + /* declare the link dead and buried */ > + ctd_target->detect_completed = > + EMCCTD_TARGET_DETECT_NOT_COMPLETED; > + memset(&ctd_target->ctd_detect, 0x0, > + sizeof(struct emc_ctd_v010_detect)); > + > + wake_up(&lun_discovery_event_barrier); > + } > + /* Connect case ... need to scan and create the needed > + * objects in the SCSI midlayer > + */ > + else { > + ctd_dprintk_crit("ctd_target -> %p\n", > + ctd_target); > + scsi_scan_target(&ctd_host->shost->shost_gendev, > + 0, i, SCAN_WILD_CARD, 1); > + lun_discovery_complete = 1; > + wake_up(&lun_discovery_event_barrier); > + } > + break; > + default: > + ctd_dprintk_crit("ctd_target -> %p detect unknown -> %x\n", > + ctd_target, ctd_target->detect_completed); > + } > + } > + > + return error; > +} > + > +/* This function posts the detect event into adapter specific list */ > +static int > +ctd_post_event(union emc_ctd_v010_message *io_msg, > + struct ctd_pci_private *ctd_private) > +{ > + int error; > + struct ctd_event_io_element *event; > + > + error = -ENOMEM; > + > + event = (struct ctd_event_io_element *) > + kzalloc(sizeof(struct ctd_event_io_element), > + GFP_ATOMIC); > + if (event) { > + error = 0; > + event->io_msg = *io_msg; > + spin_lock(&ctd_private->event_io_lock); > + list_add_tail(&event->list, &ctd_private->event_io_list); > + spin_unlock(&ctd_private->event_io_lock); > + } else { > + ctd_dprintk_crit("kzalloc failure\n"); > + } > + return error; > +} > + > +/* Thread Handler Function. This consumes the events posted into its queue, and > + * takes respective action > + */ > +static int > +ctd_event_handler(void *ctd_thread_args) > +{ > + struct ctd_event_io_element *event; > + struct ctd_pci_private *ctd_private; > + > + ctd_private = (struct ctd_pci_private *)ctd_thread_args; > + > + while (!kthread_should_stop()) { > + schedule_timeout_interruptible(HZ); > + > + event = NULL; > + > + spin_lock(&ctd_private->event_io_lock); > + if (!list_empty(&ctd_private->event_io_list)) { > + event = list_first_entry(&ctd_private->event_io_list, > + struct ctd_event_io_element, list); > + list_del(&event->list); > + } > + spin_unlock(&ctd_private->event_io_lock); > + > + > + if (event) { > + int error; > + emc_ctd_uint8_t what; > + union emc_ctd_v010_message *io_msg; > + struct emc_ctd_v010_detect *io_detect; > + > + io_msg = &event->io_msg; > + what = io_msg->emc_ctd_scsi_message_header_what; > + > + if (what != EMC_CTD_V010_WHAT_DETECT) { > + ctd_dprintk_crit("event -> %p what -> %x\n", > + event, what); > + } else { > + error = -ENODEV; > + io_detect = > + &io_msg->emc_ctd_v010_message_detect; > + > + if (io_detect->emc_ctd_v010_detect_flags == > + 0x0) { > + error = ctd_handle_disconnect( > + io_detect, ctd_private); > + } else { > + if (io_detect->emc_ctd_v010_detect_flags & > + EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET) { > + ctd_dprintk( > + "header addr -> %x key -> %llx\n", > + io_detect->emc_ctd_detect_header_address, > + io_detect->emc_ctd_v010_detect_key); > + error = ctd_handle_target_addition(io_detect, > + ctd_private); > + } > + if (io_detect->emc_ctd_v010_detect_flags & > + EMC_CTD_V010_DETECT_FLAG_SCSI_INITIATOR) { > + ctd_dprintk("\n"); > + error = ctd_handle_source_addition(io_detect, > + ctd_private); > + } > + } > + if (!error) { > + int retry = EMCCTD_DETECT_RETRY_MAX; > + > + error = ctd_ITnexus_handler(ctd_private); > + /* In case of ITnexus_handler failure, > + * pause for 2 seconds before retrying > + * the operation again > + */ > + while (error && retry) { > + schedule_timeout_interruptible(HZ * 2); > + error = ctd_ITnexus_handler(ctd_private); > + retry--; > + } while (error && retry); > + > + } > + } > + kfree(event); > + } > + } > + return 0; > +} > + > +static int > +ctd_init_event_thread(struct ctd_pci_private *ctd_private) > +{ > + int error; > + > + INIT_LIST_HEAD(&ctd_private->event_io_list); > + spin_lock_init(&ctd_private->event_io_lock); > + > + /* Create the daemon thread to handle detect requests */ > + ctd_private->ctd_event_thread = kthread_create(ctd_event_handler, > + (void *)ctd_private, "emcctd_event_thread"); > + error = (!IS_ERR(ctd_private->ctd_event_thread)) ? 0 : -EBUSY; > + if (!error) { > + wake_up_process(ctd_private->ctd_event_thread); > + } else { > + ctd_dprintk_crit( > + "FAILURE, ctd_private -> %p\n", ctd_private); > + } > + return error; > +} > + > +static void > +ctd_destroy_event_thread(struct ctd_pci_private *ctd_private) > +{ > + if (ctd_private->ctd_event_thread) > + kthread_stop(ctd_private->ctd_event_thread); > +} > + > +static void > +ctd_init_scsi_host_private(struct Scsi_Host *shost, struct pci_dev *pci_dev) > +{ > + int i; > + struct ctd_host_info *ctd_host_info; > + struct ctd_pci_private *ctd_private; > + > + ctd_private = pci_get_drvdata(pci_dev); > + > + ctd_dprintk("ctd_private -> %p\n", ctd_private); > + > + ctd_host_info = shost_priv(shost); > + memset(ctd_host_info, 0x0, sizeof(struct ctd_host_info)); > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > + struct ctd_target_info *ctd_target; > + > + ctd_target = &ctd_host_info->target[i]; > + /* nothing to do; > + * ctd_target->ctd_detect.emc_ctd_v010_detect_flags already zero > + */ > + } > + > + > + ctd_host_info->shost = shost; > + ctd_host_info->pci_dev = pci_dev; > + > + shost->can_queue = ctd_private->pci_request_queue_size; > + shost->cmd_per_lun = min(emcctd_cmd_per_lun, shost->can_queue); > + shost->max_lun = emcctd_max_luns; > + shost->max_id = EMCCTD_MAX_ID; > + > + ctd_private->host_private = ctd_host_info; > + > + ctd_dprintk("scsi_ctd_host = %p\n", ctd_host_info); > +} > + > + > +static int > +ctd_scsi_layer_init(struct pci_dev *pci_dev) > +{ > + int error; > + struct Scsi_Host *scsi_ctd_host; > + struct ctd_pci_private *ctd_private = NULL; > + > + ctd_dprintk("pci_dev -> %p\n", pci_dev); > + > + scsi_ctd_host = scsi_host_alloc(&scsi_ctd_template, > + sizeof(struct ctd_host_info)); > + if (scsi_ctd_host == NULL) { > + error = -ENOMEM; > + goto ctd_scsi_layer_init_complete; > + } > + > + ctd_init_scsi_host_private(scsi_ctd_host, pci_dev); > + > + ctd_private = pci_get_drvdata(pci_dev); > + > + error = ctd_init_event_thread(ctd_private); > + if (error) > + goto ctd_scsi_layer_init_complete; > + > + /* > + * register the HBA to the Linux SCSI stack > + */ > + error = scsi_add_host(scsi_ctd_host, &pci_dev->dev); > + > +ctd_scsi_layer_init_complete: > + if (error) { > + ctd_dprintk_crit("failure, error = %x\n", error); > + if (scsi_ctd_host != NULL) > + scsi_host_put(scsi_ctd_host); > + > + if (ctd_private) > + ctd_destroy_event_thread(ctd_private); > + } > + return error; > +} > + > +static void > +ctd_clear_io_queue(struct ctd_pci_private *ctd_private) > +{ > + struct list_head iochain; > + unsigned long flags; > + struct ctd_request_private *request, *request_next; > + > + ctd_dprintk_crit("ctd_private -> %p\n", ctd_private); > + > + INIT_LIST_HEAD(&iochain); > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + /* post reset need to cleanup the aborted io > + * as no reply is expected on them > + */ > + /* request is still kept as REPLY_AWAITED to > + * handle any response post connect > + */ > + list_for_each_entry_safe(request, request_next, > + &ctd_private->aborted_io_list, list) { > + list_del(&request->list); > + request->io_state = CTD_IO_REQUEST_REPLY_AWAITED; > + } > + > + /* rifle thru queued and requeued IO list and mark them for abort, > + * the completion to the upper layers is handled by the timeout logic > + * invoked from the SCSI midlayer > + */ > + /* request is still kept as REPLY_AWAITED to handle any response post > + * connect > + */ > + list_for_each_entry_safe(request, request_next, > + &ctd_private->queued_io_list, list) { > + list_del(&request->list); > + list_add(&request->list, &iochain); > + request->cmnd->host_scribble = NULL; > + request->io_state = CTD_IO_REQUEST_REPLY_AWAITED; > + > + /* These requests shall be aborted to upper layer, so treat them > + * as abort_sent > + */ > + atomic_long_inc(&ctd_private->hw_stats.abort_sent); > + atomic_long_dec(&ctd_private->hw_stats.active_io_count); > + } > + > + list_for_each_entry_safe(request, request_next, > + &ctd_private->requeued_io_list, list) { > + list_del(&request->list); > + list_add(&request->list, &iochain); > + request->cmnd->host_scribble = NULL; > + request->io_state = CTD_IO_REQUEST_REPLY_AWAITED; > + > + /* These requests shall be aborted to upper layer, so treat them > + * as abort_sent > + */ > + atomic_long_inc(&ctd_private->hw_stats.abort_sent); > + atomic_long_dec(&ctd_private->hw_stats.active_io_count); > + } > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + > + list_for_each_entry_safe(request, request_next, > + &iochain, list) { > + struct scsi_cmnd *cmnd; > + > + list_del(&request->list); > + > + ctd_dprintk_crit( > + "cmnd -> %p request -> %p CTD_IO_REQUEST_REPLY_AWAITED\n", > + request->cmnd, request); > + > + cmnd = request->cmnd; > + request->cmnd = NULL; > + > + /* error propagation to the SCSI midlayer */ > + scsi_translate_sam_code(cmnd, SAM_STAT_TASK_ABORTED); > + scsi_set_resid(cmnd, scsi_bufflen(cmnd)); > + cmnd->scsi_done(cmnd); > + } > + ctd_dprintk("ctd_private -> %p\n", ctd_private); > +} > + > +static int > +ctd_scsi_layer_cleanup(struct pci_dev *pci_dev) > +{ > + int error; > + struct ctd_pci_private *ctd_private; > + struct ctd_host_info *ctd_host_info; > + > + error = 0; > + > + ctd_private = pci_get_drvdata(pci_dev); > + > + ctd_dprintk("ctd_private pci_dev -> %p %p\n", ctd_private, pci_dev); > + > + ctd_check_response_queue((unsigned long)pci_dev); > + > + ctd_clear_io_queue(ctd_private); > + > + ctd_destroy_event_thread(ctd_private); > + > + flush_scheduled_work(); > + > + ctd_host_info = ctd_private->host_private; > + > + scsi_remove_host(ctd_host_info->shost); > + > + scsi_host_put(ctd_host_info->shost); > + > + return error; > +} > + > +#ifdef CONFIG_PM > +static int > +ctd_pci_suspend(struct pci_dev *pci_dev, pm_message_t state) > +{ > + pci_save_state(pci_dev); > + pci_set_power_state(pci_dev, PCI_D3hot); > + return 0; > +} > + > +static int > +ctd_pci_resume(struct pci_dev *pci_dev) > +{ > + pci_restore_state(pci_dev); > + pci_set_power_state(pci_dev, PCI_D0); > + return 0; > +} > +#endif > + > +static void > +ctd_pci_remove(struct pci_dev *pci_dev) > +{ > + struct ctd_pci_private *ctd_private; > + > + ctd_dprintk("pic_dev -> %p\n", pci_dev); > + > + ctd_private = pci_get_drvdata(pci_dev); > + > + ctd_private->hw_state = CTD_HW_STATE_DISABLED; > + > + ctd_scsi_layer_cleanup(pci_dev); > + > +#if !defined(__VMKLNX__) > + ctd_proc_remove(pci_dev); > +#endif > + free_irq(pci_dev->irq, pci_dev); > + > + pci_disable_msi(pci_dev); > + > + if (ctd_private->ioaddr_txrx_rings) > + pci_iounmap(pci_dev, ctd_private->ioaddr_txrx_rings); > + > + if (ctd_private->ioaddr_fast_registers) > + pci_iounmap(pci_dev, ctd_private->ioaddr_fast_registers); > + > + if (ctd_private->ioaddr_slow_registers) > + pci_iounmap(pci_dev, ctd_private->ioaddr_slow_registers); > + > + tasklet_kill(&ctd_private->isr_tasklet); > + > + ctd_release_io_pool(ctd_private); > + > + kfree(ctd_private); > + > + pci_release_regions(pci_dev); > + pci_set_drvdata(pci_dev, NULL); > + pci_disable_device(pci_dev); > +} > + > +static void > +ctd_check_error_condition(struct pci_dev *pci_dev) > +{ > +#define EMCCTD_MAX_CACHED_ERROR 14 > + int i, j; > + int error; > + union emc_ctd_v010_message message; > + static emc_ctd_uint32_t internal_errors_1_14[EMCCTD_MAX_CACHED_ERROR]; > + struct ctd_pci_private *ctd_private = pci_get_drvdata(pci_dev); > + > + if (ctd_private->pci_fast_registers->emc_ctd_v010_fregs_error_flag) { > + for (i = 0; i < EMCCTD_MAX_CACHED_ERROR; i++) { > + if (internal_errors_1_14[i] != > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_errors_1_14[i]) { > + > + internal_errors_1_14[i] = > + ctd_private->pci_fast_registers->emc_ctd_v010_fregs_errors_1_14[i]; > + > + error = i + 1; > + > + for (j = 0; j < EMC_CTD_V010_LOG_ERROR_TX_SIZE; j++) { > + if (ctd_private->pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_error[j] == > + error) { > + memcpy(&message, > + &ctd_private->pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_message[j], > + sizeof(message)); > + ctd_dprintk_crit( > + "header addr -> %x error -> %s\n", > + message.emc_ctd_scsi_message_header_address, > + (error == > + EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED ? > + "EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED" : > + error == > + EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT ? > + "EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT" : > + error == > + EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED ? > + "EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED" : > + error == > + EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER ? > + "EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER" : > + error == > + EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE ? > + "EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE" : > + error == > + EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN ? > + "EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN" : > + "EMC_CTD_V010_ERROR_NULL")); > + } > + } > + } > + } > + } > +} Ugh... What about: --- 8< --- if (!ctd_private->pci_fast_registers->emc_ctd_v010_fregs_error_flag) return; for (i = 0; i < EMCCTD_MAX_CACHED_ERROR; i++) { if (internal_errors_1_14[i] == ctd_private->pci_fast_registers->emc_ctd_v010_fregs_errors_1_14[i]) continue; internal_errors_1_14[i] = ctd_private->pci_fast_registers->emc_ctd_v010_fregs_errors_1_14[i]; error = i + 1; for (j = 0; j < EMC_CTD_V010_LOG_ERROR_TX_SIZE; j++) { if (ctd_private->pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_error[j] != error) continue; memcpy(&message, &ctd_private->pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_message[j], sizeof(message)); ctd_dprintk_crit( "header addr -> %x error -> %s\n", message.emc_ctd_scsi_message_header_address, (error == EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED ? "EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED" : error == EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT ? "EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT" : error == EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED ? "EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED" : error == EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER ? "EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER" : error == EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE ? "EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE" : error == EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN ? "EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN" : "EMC_CTD_V010_ERROR_NULL")); } } --- >8 --- And that's only to save indedation levels. This whole part can be made something like that: char *msg[] = { [EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED] = "EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED", [EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT] = "EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT", [EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED] = "EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED", [EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER] = "EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER", [EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE] = "EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE", [EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN] = "EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN", [EMC_CTD_V010_ERROR_NULL] = "EMC_CTD_V010_ERROR_NULL", } You can probably even use __stringify() for that. > +/* > + * ctd_check_response_queue > + * > + * Bottom half of interrupt handler. > + */ > +static void > +ctd_check_response_queue(unsigned long instance_addr) > +{ > + struct pci_dev *pci_dev = (struct pci_dev *)instance_addr; > + struct ctd_pci_private *ctd_private = pci_get_drvdata(pci_dev); > + union emc_ctd_v010_message io_response; > + > + > + /* empty response queue */ > + while (ctd_hw_dequeue_response(&io_response, ctd_private) == > + SUCCESS) { > + /* handle the response */ > + ctd_handle_response(&io_response, ctd_private); > + } > +} Wrong indendt here. > + > + > +static irqreturn_t > +ctd_isr(int irq, void *opaque) > +{ > + struct pci_dev *pci_dev = (struct pci_dev *)opaque; > + struct ctd_pci_private *ctd_private = pci_get_drvdata(pci_dev); > + > + atomic_long_inc(&ctd_private->hw_stats.interrupts); > + > + /* schedule work for later */ > + tasklet_schedule(&ctd_private->isr_tasklet); > + > + return IRQ_HANDLED; > +} > + > +static int > +ctd_request_msi(struct pci_dev *pci_dev) > +{ > + int err = -EFAULT; > + > + if (pci_dev->irq) { > + err = pci_enable_msi(pci_dev); > + if (!err) { > + err = request_irq(pci_dev->irq, ctd_isr, > + IRQF_SHARED, > + pci_name(pci_dev), pci_dev); > + if (err < 0) { > + ctd_dprintk_crit("request_irq failure !!!\n"); > + pci_disable_msi(pci_dev); > + err = -EBUSY; > + } > + } > + } > + return err; > +} > + > +static struct ctd_request_private * > +ctd_acquire_request(struct ctd_pci_private *ctd_private) > +{ > + unsigned long flags; > + struct ctd_request_private *ctd_request; > + > + ctd_request = NULL; > + > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + > + /* check if any request in the aborted io list can be reused */ > + if (!list_empty(&ctd_private->aborted_io_list)) { > + struct ctd_request_private *request, *request_next; > + > + list_for_each_entry_safe(request, request_next, > + &ctd_private->aborted_io_list, list) { > + /* aborted_io_list is in chronologically order thus > + * failure of time_after() indicates any request after > + * this point is not in the kill zone > + */ > + if (time_before(jiffies, request->purge_lifetime)) > + break; > + > + list_del(&request->list); > + if (request->cdb_page) { > + __free_pages(request->cdb_page, > + request->cdb_page_order); > + } > + if (request->sgllist_page) { > + __free_pages(request->sgllist_page, > + request->sgllist_page_order); > + } > + memset(request, 0x0, > + sizeof(struct ctd_request_private)); > + list_add(&request->list, &ctd_private->io_pool); > + atomic_long_inc(&ctd_private->hw_stats.free_io_entries); > + } > + } > + > + if (!list_empty(&ctd_private->io_pool)) { > + ctd_request = list_first_entry(&ctd_private->io_pool, > + struct ctd_request_private, list); > + list_del(&ctd_request->list); > + ctd_request->io_state = CTD_IO_REQUEST_QUEUED; > + } > + > + if (ctd_request) > + atomic_long_dec(&ctd_private->hw_stats.free_io_entries); > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > + > + return ctd_request; > +} > + > +static void > +ctd_release_request(struct ctd_request_private *ctd_request, > + struct ctd_pci_private *ctd_private) > +{ > + unsigned long flags; > + > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > + memset(ctd_request, 0x0, sizeof(struct ctd_request_private)); > + list_add(&ctd_request->list, &ctd_private->io_pool); > + > + atomic_long_inc(&ctd_private->hw_stats.free_io_entries); > + > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > +} > + > +static void > +ctd_release_io_pool(struct ctd_pci_private *ctd_private) > +{ > + kfree(ctd_private->io_map); > +} > + > +static int > +ctd_alloc_io_pool(struct ctd_pci_private *ctd_private, unsigned int pool_size) > +{ > + int i; > + int error; > + > + error = -ENOMEM; > + > + INIT_LIST_HEAD(&ctd_private->io_pool); > + INIT_LIST_HEAD(&ctd_private->queued_io_list); > + INIT_LIST_HEAD(&ctd_private->aborted_io_list); > + INIT_LIST_HEAD(&ctd_private->requeued_io_list); > + > + spin_lock_init(&ctd_private->io_mgmt_lock); > + > + ctd_private->io_map = kcalloc(pool_size, > + sizeof(struct ctd_request_private), GFP_KERNEL); > + > + /* > + * in case of allocation failure try with one fourth the size before > + * throwing the towel > + */ > + > + if (ctd_private->io_map == NULL) { > + pool_size = pool_size >> 2; > + ctd_private->io_map = kcalloc(pool_size, > + sizeof(struct ctd_request_private), GFP_KERNEL); > + if (ctd_private->io_map == NULL) { > + ctd_dprintk_crit( > + "io_pool allocation failure for pool_size -> %d\n", > + pool_size); > + goto ctd_alloc_io_pool_complete; > + } > + } > + > + ctd_private->io_map_end = ctd_private->io_map + pool_size; > + > + for (i = 0; i < pool_size; i++) { > + struct ctd_request_private *request_context = > + ctd_private->io_map + i; > + memset(request_context, 0x0, > + sizeof(struct ctd_request_private)); > + list_add(&request_context->list, &ctd_private->io_pool); > + } > + ctd_dprintk_crit( > + "ctd_private -> %p, pool_size -> %x, io_map -> %p, io_map_end-> %p\n", > + ctd_private, pool_size, > + ctd_private->io_map, ctd_private->io_map_end); > + error = 0; > + ctd_private->hw_stats.free_io_entries.counter = pool_size; > +ctd_alloc_io_pool_complete: > + return error; > +} > + > +static int > +ctd_pci_probe( > + struct pci_dev *pci_dev, > + const struct pci_device_id *id > +) > +{ > + struct ctd_pci_private *ctd_private; > + int ctd_proc_initialized; > + int ctd_scsi_initialized; > + int ctd_regions_initialized; > + int err; > + > + err = -ENODEV; > + ctd_proc_initialized = FAILED; > + ctd_scsi_initialized = FAILED; > + ctd_regions_initialized = FAILED; > + > + ctd_private = (struct ctd_pci_private *) > + kzalloc(sizeof(struct ctd_pci_private), GFP_ATOMIC); > + > + if (ctd_private == NULL) { > + ctd_dprintk_crit("kzalloc Failure\n"); > + goto ctd_pci_probe_complete; > + } > + > + ctd_private->pci_dev = pci_dev; > + > + /* enable the device */ > + err = pci_enable_device(pci_dev); > + if (err) { > + ctd_dprintk_crit("pci_enable_device Failure\n"); > + goto ctd_pci_probe_complete; > + } > + pci_set_master(pci_dev); > + > + err = pci_request_regions(pci_dev, "ctd-pci"); > + if (err) { > + ctd_dprintk_crit("pci_request_regions Failure\n"); > + goto ctd_pci_probe_complete; > + } > + > + ctd_regions_initialized = SUCCESS; > + > + ctd_dprintk("ctd_private pci_dev -> %p %p\n", ctd_private, pci_dev); > + > +#define EMC_CTD_TXRX_MSG_SIZE 128 > + > + if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_RINGS)) { > + ctd_private->ioaddr_txrx_rings = ioremap( > + pci_resource_start(pci_dev, > + EMC_CTD_V010_BAR_RINGS), > + pci_resource_len(pci_dev, > + EMC_CTD_V010_BAR_RINGS) > + ); > + ctd_private->txrx_ringsize = > + (pci_resource_len(pci_dev, > + EMC_CTD_V010_BAR_RINGS) >> 1) > + / EMC_CTD_TXRX_MSG_SIZE; > + > + ctd_dprintk_crit( > + "physical addr = %llx ioaddr_txrx_rings = %p , ring size = %x\n", > + pci_resource_start(pci_dev, > + EMC_CTD_V010_BAR_RINGS), > + ctd_private->ioaddr_txrx_rings, > + ctd_private->txrx_ringsize); > + > + } > + if (ctd_private->ioaddr_txrx_rings == NULL) { > + err = -ENOMEM; > + ctd_dprintk_crit("ioremap failure\n"); > + goto ctd_pci_probe_complete; > + } else { > + ctd_private->pci_request_array = ctd_private->ioaddr_txrx_rings; > + ctd_private->pci_response_array = > + ctd_private->ioaddr_txrx_rings + > + ((pci_resource_len(pci_dev, > + EMC_CTD_V010_BAR_RINGS)) >> 1); > + } > + > + > + if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_FREGS)) { > + ctd_private->ioaddr_fast_registers = > + ioremap(pci_resource_start(pci_dev, > + EMC_CTD_V010_BAR_FREGS), > + pci_resource_len(pci_dev, > + EMC_CTD_V010_BAR_FREGS)); ioremap() can fail. > + > + ctd_dprintk_crit( > + "physical addr = %llx ioaddr_fast_registers = %p\n", > + pci_resource_start(pci_dev, > + EMC_CTD_V010_BAR_FREGS), > + ctd_private->ioaddr_fast_registers); > + } > + if (ctd_private->ioaddr_fast_registers == NULL) { > + err = -ENOMEM; > + ctd_dprintk_crit("ioremap failure\n"); > + goto ctd_pci_probe_complete; > + } else { > + ctd_private->pci_fast_registers = > + ctd_private->ioaddr_fast_registers; > + } > + > + if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_SREGS)) { > + ctd_private->ioaddr_slow_registers = > + ioremap(pci_resource_start(pci_dev, > + EMC_CTD_V010_BAR_SREGS), > + pci_resource_len(pci_dev, > + EMC_CTD_V010_BAR_SREGS)); > + Same goes here. > + ctd_dprintk_crit( > + "physical addr = %llx ioaddr_slow_registers = %p\n", > + pci_resource_start(pci_dev, > + EMC_CTD_V010_BAR_SREGS), > + ctd_private->ioaddr_slow_registers); > + } > + if (ctd_private->ioaddr_slow_registers == NULL) { > + err = -ENOMEM; > + ctd_dprintk_crit("ioremap failure\n"); > + goto ctd_pci_probe_complete; > + } else { > + ctd_private->pci_slow_registers = > + ctd_private->ioaddr_slow_registers; > + } > + > + /* reset the device */ > + ctd_private->pci_device_reset_register = 0XFF; > + > + ctd_private->pci_request_queue_size = ctd_private->txrx_ringsize; > + ctd_private->pci_response_queue_size = ctd_private->txrx_ringsize; > + > + err = ctd_alloc_io_pool(ctd_private, > + ctd_private->pci_request_queue_size); > + if (err) { > + ctd_dprintk_crit("ctd_alloc_io_pool failure\n"); > + goto ctd_pci_probe_complete; > + } > + > + pci_set_drvdata(pci_dev, ctd_private); > + > + spin_lock_init(&ctd_private->isr_lock); > + > + /* setup tasklet for scanning response queue */ > + tasklet_init(&ctd_private->isr_tasklet, > + ctd_check_response_queue, (unsigned long)pci_dev); > + > + ctd_private->hw_state = CTD_HW_STATE_INITIALIZED; > + > + pci_set_master(pci_dev); > + See above. > +#if !defined(__VMKLNX__) > + err = ctd_proc_init(pci_dev); > + if (err) { > + ctd_dprintk_crit("ctd_proc_init failure\n"); > + goto ctd_pci_probe_complete; > + } > + ctd_proc_initialized = SUCCESS; > +#endif > + err = ctd_scsi_layer_init(pci_dev); > + if (err) { > + ctd_dprintk_crit("ctd_scsi_layer_init failure\n"); > + goto ctd_pci_probe_complete; > + } > + ctd_scsi_initialized = SUCCESS; > + > + err = ctd_request_msi(pci_dev); > + if (err) { > + ctd_dprintk_crit("ctd_request_msi failure\n"); > + goto ctd_pci_probe_complete; > + } > + > + /* after we reset the device, but before we enabled MSI, some messages > + * may have been received. check for them > + */ > + tasklet_schedule(&ctd_private->isr_tasklet); > + > +ctd_pci_probe_complete: > + if (err) { > + if (ctd_private) { > + tasklet_kill(&ctd_private->isr_tasklet); > + > + if (ctd_scsi_initialized == SUCCESS) > + ctd_scsi_layer_cleanup(pci_dev); > + > + if (ctd_proc_initialized == SUCCESS) > + ctd_proc_remove(pci_dev); > + > + if (ctd_private->ioaddr_txrx_rings) > + pci_iounmap(pci_dev, > + ctd_private->ioaddr_txrx_rings); > + > + if (ctd_private->ioaddr_fast_registers) > + pci_iounmap(pci_dev, > + ctd_private->ioaddr_fast_registers); > + > + if (ctd_private->ioaddr_slow_registers) > + pci_iounmap(pci_dev, > + ctd_private->ioaddr_slow_registers); > + > + if (ctd_regions_initialized == SUCCESS) > + pci_release_regions(pci_dev); > + > + ctd_release_io_pool(ctd_private); > + kfree(ctd_private); > + } > + pci_set_drvdata(pci_dev, NULL); > + pci_disable_device(pci_dev); > + } > + return err; > +} > + No proc please. > +#if !defined(__VMKLNX__) > +static const struct file_operations ctd_proc_fops = { > + .open = ctd_proc_open, > + .read = seq_read, > + .llseek = seq_lseek, > + .release = seq_release, > +}; > +static int > +ctd_proc_open(struct inode *inode, struct file *file) > +{ > + return single_open(file, ctd_proc_show, PDE_DATA(inode)); > +} > + > +int > +ctd_proc_init(struct pci_dev *pci_dev) > +{ > + int err; > + static int hw_index; > + char hw_name[MAX_PROC_FILE_NAMELEN]; > + struct ctd_pci_private *ctd_private; > + struct proc_dir_entry *pde; > + > + ctd_private = (struct ctd_pci_private *) pci_get_drvdata(pci_dev); > + ctd_private->hw_index = hw_index; > + > + memset(hw_name, 0x0, sizeof(hw_name)); > + snprintf(hw_name, sizeof(hw_name), "emcctd_stats_%d", hw_index++); > + > + err = -EPERM; > + > + do { > + if (ctd_proc_directory == NULL) > + break; > + pde = proc_create_data(hw_name, S_IFREG | S_IRUGO | S_IWUSR, > + ctd_proc_directory, > + &ctd_proc_fops, > + ctd_private); > + > + if (pde == NULL) { > + ctd_dprintk_crit( > + "create_proc_read_entry failure for %s\n", > + hw_name); > + break; > + } > + err = 0; > + } while (0); > + return err; > +} > + > +void > +ctd_proc_remove(struct pci_dev *pci_dev) > +{ > + int hw_index; > + char hw_name[MAX_PROC_FILE_NAMELEN]; > + struct ctd_pci_private *ctd_private; > + > + ctd_private = (struct ctd_pci_private *) pci_get_drvdata(pci_dev); > + > + hw_index = ctd_private->hw_index; > + memset(hw_name, 0x0, sizeof(hw_name)); > + snprintf(hw_name, sizeof(hw_name), "emc/emcctd_stats_%d", hw_index); > + > + ctd_dprintk("removing %s\n", hw_name); > + > + remove_proc_entry(hw_name, NULL); > + > +} > +#endif > + > +static int __init ctd_pci_init(void) > +{ > + int err; > + > + ctd_dprintk_crit("Loading emcctd\n"); > + init_waitqueue_head(&lun_discovery_event_barrier); > + > + ctd_proc_directory = proc_mkdir("emc", NULL); > + > + err = pci_register_driver(&ctd_pci_driver); > + > + if (err) { > + remove_proc_entry("emc", NULL); > + } else { > + /* wait for 20 seconds or less to allow the luns to appear > + * before exiting from insmod > + */ > + wait_event_interruptible_timeout(lun_discovery_event_barrier, > + lun_discovery_complete, ((HZ) * 20)); > + } > + > + return err; > +} > + > +module_init(ctd_pci_init); > + > +static void __exit ctd_pci_exit(void) > +{ > + pci_unregister_driver(&ctd_pci_driver); > + remove_proc_entry("emc", NULL); > +} > + > +module_exit(ctd_pci_exit); > diff --git a/drivers/scsi/emcctd/emcctd.h b/drivers/scsi/emcctd/emcctd.h > new file mode 100644 > index 0000000..3fba931 > --- /dev/null > +++ b/drivers/scsi/emcctd/emcctd.h > @@ -0,0 +1,232 @@ > +/* > + * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem. > + * > + * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA. > + * > + * Authors: > + * fredette, matt <matt.fredette@xxxxxxx> > + * Pirotte, Serge <serge.pirotte@xxxxxxx> > + * Singh Animesh <Animesh.Singh@xxxxxxx> > + * Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > + * > + * This program is free software; you can redistribute it and/or > + * modify it under the terms of the GNU General Public License > + * as published by the Free Software Foundation; either version 2 > + * of the License, or (at your option) any later version. > + * > + * This program is distributed in the hope that it will be useful, > + * but WITHOUT ANY WARRANTY; without even the implied warranty of > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the > + * GNU General Public License for more details. > + * > + */ > + > +#ifndef _EMCCTD_H_ > +#define _EMCCTD_H_ > + > +#define DRV_NAME "emcctd" > + > +#define EMCCTD_V010_PROTOCOL_MINOR_VERSION 0x0 > + > +/* please refer to emc_ctd_v010_scsi_command_cdb in emc_ctd_interface.h */ > +#define EMCCTD_V010_MAX_CDB_SIZE 16 > + > +#define EMCCTD_MAX_LUN 16384 > +#define EMCCTD_MAX_ID 16 > +#define EMCCTD_MAX_RETRY 5 > +#define EMCCTD_CMD_PER_LUN 16 > +#define EMCCTD_THIS_ID (-1) > +#define EMCCTD_REQUEST_TIMEOUT (60 * HZ) > +#define EMCCTD_OPAQUE_PURGE_WAITTIME (10 * HZ) > + > +#define EMCCTD_DEVICE_RESET_PAUSE 3 > +#define EMCCTD_DETECT_RETRY_MAX 3 > + > +#define ctd_dprintk(__m_fmt, ...) \ > +do { \ > + if (ctd_debug) \ > + pr_info("%s:%d:"__m_fmt, __func__, __LINE__, ##__VA_ARGS__); \ > +} while (0) Please use pr_debug() here. > + > +#define ctd_dprintk_crit(__m_fmt, ...) \ > + pr_crit("%s:%d:"__m_fmt, __func__, __LINE__, ##__VA_ARGS__) File and line information is probably not of any interest for the users and serves a debugging purpose only. > + > +#define EMCCTD_TARGET_DETECT_COMPLETED 1 > +#define EMCCTD_TARGET_DETECT_NOT_COMPLETED 0 > + > +struct ctd_target_info { > + unsigned int detect_completed; > + struct scsi_target *starget; > + struct emc_ctd_v010_detect ctd_detect; > +}; > + > +#define emc_ctd_detect_header_address \ > + emc_ctd_v010_detect_header.emc_ctd_v010_header_address > +#define emc_ctd_detect_header_minor \ > + emc_ctd_v010_detect_header.emc_ctd_v010_header_minor > +#define emc_ctd_detect_header_what \ > + emc_ctd_v010_detect_header.emc_ctd_v010_header_what > + > +#define emc_ctd_scsi_command_header_address \ > + emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_address > +#define emc_ctd_scsi_command_header_minor \ > + emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_minor > +#define emc_ctd_scsi_command_header_what \ > + emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_what > + > +#define emc_ctd_scsi_response_header_address \ > + emc_ctd_v010_scsi_response_header.emc_ctd_v010_header_address > +#define emc_ctd_scsi_response_header_minor \ > + emc_ctd_v010_scsi_response_header.emc_ctd_v010_header_minor > +#define emc_ctd_scsi_response_header_what \ > + emc_ctd_v010_scsi_response_header.emc_ctd_v010_header_what > + > +#define emc_ctd_scsi_phase_header_address \ > + emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_address > +#define emc_ctd_scsi_phase_header_minor \ > + emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_minor > +#define emc_ctd_scsi_phase_header_what \ > + emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_what > + > +#define emc_ctd_scsi_message_header_address \ > + emc_ctd_v010_message_header.emc_ctd_v010_header_address > +#define emc_ctd_scsi_message_header_minor \ > + emc_ctd_v010_message_header.emc_ctd_v010_header_minor > +#define emc_ctd_scsi_message_header_what \ > + emc_ctd_v010_message_header.emc_ctd_v010_header_what > + > +#define emc_ctd_scsi_command_sgl \ > + emc_ctd_v010_scsi_command_u.emc_ctd_v010_scsi_command_sgl > +#define emc_ctd_scsi_response_extra \ > + emc_ctd_v010_scsi_response_u.emc_ctd_v010_scsi_response_extra > + > +#define ctd_detect_name_bytes \ > + ctd_detect.emc_ctd_v010_detect_name.emc_ctd_v010_name_bytes > + > +struct ctd_host_info { > + struct Scsi_Host *shost; > + struct pci_dev *pci_dev; > + struct ctd_target_info target[EMCCTD_MAX_ID]; > +}; > + > +struct ctd_dev_info { > + struct ctd_host_info *ctd_host; > + struct ctd_target_info *ctd_target; > + struct emc_ctd_v010_detect *ctd_target_detect; > +}; > + > +#define PROC_STAT_SCSI_TS_MAX 10 > +#define MAX_PROC_FILE_NAMELEN 128 > +#define CTD_MAX_IO_STATS 200 > + > +struct ctd_hw_stats { > + atomic_long_t interrupts; > + atomic_long_t requests_sent; > + atomic_long_t responses_received; > + atomic_long_t active_io_count; > + atomic_long_t abort_sent; > + atomic_long_t abort_received; > + atomic_long_t what_in; > + atomic_long_t what_out; > + atomic_long_t free_io_entries; > + unsigned long long io_stats[CTD_MAX_IO_STATS]; > + unsigned int io_stats_index; > +}; > + > +enum ctd_io_request_state { > + CTD_IO_REQUEST_FREE, > + CTD_IO_REQUEST_QUEUED, > + CTD_IO_REQUEST_REQUEUED, > + CTD_IO_REQUEST_ABORTED, > + CTD_IO_REQUEST_COMPLETED, > + CTD_IO_REQUEST_REPLY_AWAITED, > + CTD_IO_REQUEST_INVALID > +}; > + > +enum ctd_hw_state { > + CTD_HW_STATE_UNINITIALIZED, > + CTD_HW_STATE_INITIALIZED, > + CTD_HW_STATE_UNDER_RESET, > + CTD_HW_STATE_DISABLED, > + CTD_HW_STATE_INVALID > +}; > + > +struct ctd_request_private { > + struct list_head list; > + enum ctd_io_request_state io_requeue_state; > + unsigned int io_timeout; > + enum ctd_io_request_state io_state; > + struct scsi_cmnd *cmnd; > + struct page *cdb_page; > + unsigned int cdb_page_order; > + struct page *sgllist_page; > + unsigned int sgllist_page_order; > + unsigned long purge_lifetime; > + unsigned long long io_start_time; > +}; > + > +struct ctd_pci_private { > + struct pci_dev *pci_dev; > + void *host_private; > + > + void __iomem *ioaddr_txrx_rings; > + void __iomem *ioaddr_fast_registers; > + void __iomem *ioaddr_slow_registers; > + > + emc_ctd_uint32_t txrx_ringsize; > + emc_ctd_uint32_t pci_request_queue_size; > + emc_ctd_uint32_t pci_response_queue_size; > + struct emc_ctd_v010_fregs *pci_fast_registers; > + struct emc_ctd_v010_sregs *pci_slow_registers; > + union emc_ctd_v010_message *pci_response_array; > + union emc_ctd_v010_message *pci_request_array; > + > + struct tasklet_struct isr_tasklet; > + spinlock_t isr_lock; > + unsigned int hw_index; > + struct ctd_hw_stats hw_stats; > + enum ctd_hw_state hw_state; > + struct list_head queued_io_list; > + struct list_head aborted_io_list; > + struct list_head requeued_io_list; > + struct list_head io_pool; > + struct ctd_request_private *io_map; > + struct ctd_request_private *io_map_end; > + spinlock_t io_mgmt_lock; > + > + struct task_struct *ctd_event_thread; > + struct list_head event_io_list; > + spinlock_t event_io_lock; > +}; > + > +#define request_producer_index \ > + pci_fast_registers->emc_ctd_v010_fregs_tx_index_producer > +#define request_consumer_index \ > + pci_fast_registers->emc_ctd_v010_fregs_tx_index_consumer > +#define response_producer_index \ > + pci_fast_registers->emc_ctd_v010_fregs_rx_index_producer > +#define response_consumer_index \ > + pci_fast_registers->emc_ctd_v010_fregs_rx_index_consumer > +#define pci_device_reset_register \ > + pci_slow_registers->emc_ctd_v010_sregs_reset > + > +#define pci_device_name_bytes \ > + pci_fast_registers->emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes > + > +struct ctd_event_io_element { > + struct list_head list; > + union emc_ctd_v010_message io_msg; > +}; > + Though your driver depends on X86 this is not really protable. From what I saw in the code you're using it just for time differences. Please have a look at the kernel's timer infrastructure. > +static inline unsigned long long ctd_read_tsc(void) > +{ > + unsigned long long local_tsc; > + > + local_tsc = rdtsc(); > + > + return local_tsc; > +} > + > +#endif /* _EMCCTD_H_ */ > + > +/* vi: set ts=8 sw=8 noet: */ > -- > 1.8.5.2 > > -- > To unsubscribe from this list: send the line "unsubscribe linux-scsi" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Johannes Thumshirn Storage jthumshirn@xxxxxxx +49 911 74053 689 SUSE LINUX GmbH, Maxfeldstr. 5, 90409 Nürnberg GF: Felix Imendörffer, Jane Smithard, Graham Norton HRB 21284 (AG Nürnberg) Key fingerprint = EC38 9CAB C2C4 F25D 8600 D0D0 0393 969D 2D76 0850 -- To unsubscribe from this list: send the line "unsubscribe linux-scsi" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html