Thanks for your time. My replies inlined... > -----Original Message----- > From: Johannes Thumshirn [mailto:jthumshirn@xxxxxxx] > Sent: Tuesday, January 19, 2016 9:34 PM > To: Singhal, Maneesh > Cc: linux-scsi@xxxxxxxxxxxxxxx; linux-kernel@xxxxxxxxxxxxxxx; > JBottomley@xxxxxxxx; martin.petersen@xxxxxxxxxx; linux- > api@xxxxxxxxxxxxxxx > Subject: Re: [PATCH] drivers/scsi/emcctd: drivers/scsi/emcctd: Client > driver implementation for EMC-Symmetrix GuestOS emulated Cut- > Through Device > > On Tue, Jan 19, 2016 at 11:58:06AM +0000, Singhal, Maneesh wrote: > > Hello, > > Kindly review the following patch for the following driver to be > added in SCSI subsystem - > > > > Regards > > Maneesh > > > > ---------------------------------------------------------------------------- > > From f3c4b836d6f130b1d7ded618002c8164f8f4a06d Mon Sep 17 > 00:00:00 2001 > > From: "maneesh.singhal" <maneesh.singhal@xxxxxxx> > > Date: Tue, 19 Jan 2016 06:39:35 -0500 > > Subject: [PATCH] [PATCH] drivers/scsi/emcctd: Client driver > implementation for > > EMC-Symmetrix GuestOS emulated Cut-Through Device. > > > > The patch is a driver implementation EMC-Symmetrix GuestOS > emulated Cut-Through > > Device. The Cut-Through Device PCI emulation is implemented for > GuestOS > > environments in the HyperMax OS. GuestOS environments allows > loading of > > any x86 compliant operating systems like Linux/FreeBSD etc. > > > > The client driver is a SCSI HBA implementation which interfaces with > SCSI > > midlayer in the north-bound interfaces and connects with the > emulated PCI device > > on the south side. > > > > The PCI vendor ID:product ID for emulated Cut-Through Device is > 0x1120:0x1B00. > > > > Signed-off-by: maneesh.singhal <maneesh.singhal@xxxxxxx> > > --- > > Documentation/scsi/emcctd.txt | 57 + > > MAINTAINERS | 9 + > > drivers/scsi/Kconfig | 1 + > > drivers/scsi/Makefile | 1 + > > drivers/scsi/emcctd/Kconfig | 7 + > > drivers/scsi/emcctd/Makefile | 1 + > > drivers/scsi/emcctd/README | 10 + > > drivers/scsi/emcctd/emc_ctd_interface.h | 386 +++++ > > drivers/scsi/emcctd/emcctd.c | 2840 > +++++++++++++++++++++++++++++++ > > drivers/scsi/emcctd/emcctd.h | 232 +++ > > 10 files changed, 3544 insertions(+) > > create mode 100644 Documentation/scsi/emcctd.txt > > create mode 100644 drivers/scsi/emcctd/Kconfig > > create mode 100644 drivers/scsi/emcctd/Makefile > > create mode 100644 drivers/scsi/emcctd/README > > create mode 100644 drivers/scsi/emcctd/emc_ctd_interface.h > > create mode 100644 drivers/scsi/emcctd/emcctd.c > > create mode 100644 drivers/scsi/emcctd/emcctd.h > > > > diff --git a/Documentation/scsi/emcctd.txt > b/Documentation/scsi/emcctd.txt > > new file mode 100644 > > index 0000000..bcafc87 > > --- /dev/null > > +++ b/Documentation/scsi/emcctd.txt > > @@ -0,0 +1,56 @@ > > +This file contains brief information about the EMC Cut-Through > Driver (emcctd). > > +The driver is currently maintained by Singhal, Maneesh > (maneesh.singhal@xxxxxxx) > > + > > +Last modified: Mon Jan 18 2016 by Maneesh Singhal > > + > > +BASICS > > + > > +Its a client driver implementation for EMC-Symmetrix GuestOS > emulated > > +Cut-Through Device. The Cut-Through Device PCI emulation is > implemented for > > +GuestOS environments in the HyperMax OS. GuestOS > environments allows loading of > > +any x86 compliant operating systems like Linux/FreeBSD etc. > > + > > +The client driver is a SCSI HBA implementation which interfaces with > SCSI > > +midlayer in the north-bound interfaces and connects with the > emulated PCI device > > +on the south side. > > + > > +The PCI vendor ID:product ID for emulated Cut-Through Device is > 0x1120:0x1B00. > > + > > +VERSIONING > > + > > +The Version of the driver is maintained as 2.0.0.X, where 2 refers to > the CTD > > +protocol in use, and X refers to the ongoing version of the driver. > > + > > + > > +SYSFS SUPPORT > > + > > +The driver creates the directory /sys/module/emcctd and > populates it with > > +version file and a directory for various parameters as described in > MODULE > > +PARAMETERS section. > > + > > +PROCFS SUPPORT > > + > > +The driver creates the directory /proc/emc and creates files > emcctd_stats_x > > +where 'x' refers to the PCI emulation number this client driver > connected to. > > +These files cotains WWN information and IO statistics for the > particular PCI > > +emulation. > > + > > +MODULE PARAMETERS > > + > > +The supported parameters which could add debuggability or change > the runtime > > +behavior of the driver are as following: > > + > > +ctd_debug=0 | 1 Enable driver debug messages(0=off, > 1=on) > > + > > +max_luns=xx Specify the maximum number of LUN's > per > > + host(default=16384) > > + > > +cmd_per_lun=xx Specify the maximum commands per > lun(default=16) > > + > > +DEBUGGING HINTS > > + > > +Debugging code is now compiled in by default but debugging is > turned off > > +with the kernel module parameter debug_flag defaulting to 0. To > enable debug at > > +module load time add ctd_debug=1 to the module load options. > > +Debugging can also be enabled and disabled by writing a '0' (disable) > or '1' > > +(enable) to the sysfs file > /sys/module/emcctd/parameters/ctd_debug > > diff --git a/MAINTAINERS b/MAINTAINERS > > index 1d23f70..671173a 100644 > > --- a/MAINTAINERS > > +++ b/MAINTAINERS > > @@ -6601,6 +6601,15 @@ F: drivers/message/fusion/ > > F: drivers/scsi/mpt2sas/ > > F: drivers/scsi/mpt3sas/ > > > > +EMC Cut-Through Driver > > +M: fredette, matt <matt.fredette@xxxxxxx> > > +M: Pirotte, Serge <serge.pirotte@xxxxxxx> > > +M: Singh Animesh <Animesh.Singh@xxxxxxx> > > +M: Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > > +L: linux-scsi@xxxxxxxxxxxxxxx > > +S: Maintained > > +F: drivers/scsis/emcctd/ > > + > > LSILOGIC/SYMBIOS/NCR 53C8XX and 53C1010 PCI-SCSI drivers > > M: Matthew Wilcox <matthew@xxxxxx> > > L: linux-scsi@xxxxxxxxxxxxxxx > > diff --git a/drivers/scsi/Kconfig b/drivers/scsi/Kconfig > > index c1fe0d2..cbc03d7 100644 > > --- a/drivers/scsi/Kconfig > > +++ b/drivers/scsi/Kconfig > > @@ -477,6 +477,7 @@ source "drivers/scsi/aic7xxx/Kconfig.aic79xx" > > source "drivers/scsi/aic94xx/Kconfig" > > source "drivers/scsi/hisi_sas/Kconfig" > > source "drivers/scsi/mvsas/Kconfig" > > +source "drivers/scsi/emcctd/Kconfig" > > > > config SCSI_MVUMI > > tristate "Marvell UMI driver" > > diff --git a/drivers/scsi/Makefile b/drivers/scsi/Makefile > > index 862ab4e..55bea65 100644 > > --- a/drivers/scsi/Makefile > > +++ b/drivers/scsi/Makefile > > @@ -132,6 +132,7 @@ obj-$(CONFIG_SCSI_IBMVFC) += ibmvscsi/ > > obj-$(CONFIG_SCSI_HPTIOP) += hptiop.o > > obj-$(CONFIG_SCSI_STEX) += stex.o > > obj-$(CONFIG_SCSI_MVSAS) += mvsas/ > > +obj-$(CONFIG_SCSI_EMCCTD) += emcctd/ > > obj-$(CONFIG_SCSI_MVUMI) += mvumi.o > > obj-$(CONFIG_PS3_ROM) += ps3rom.o > > obj-$(CONFIG_SCSI_CXGB3_ISCSI) += libiscsi.o libiscsi_tcp.o cxgbi/ > > diff --git a/drivers/scsi/emcctd/Kconfig > b/drivers/scsi/emcctd/Kconfig > > new file mode 100644 > > index 0000000..d995b53 > > --- /dev/null > > +++ b/drivers/scsi/emcctd/Kconfig > > @@ -0,0 +1,7 @@ > > +config SCSI_EMCCTD > > + tristate "EMC Cut-Through client driver" > > + depends on SCSI && PCI && X86 > > + help > > + This driver supports EMC virtual PCI Cut-Through- > Device. > > + To compile this driver as a module, choose M here: the > > + module will be called emcctd. > > diff --git a/drivers/scsi/emcctd/Makefile > b/drivers/scsi/emcctd/Makefile > > new file mode 100644 > > index 0000000..74a070e > > --- /dev/null > > +++ b/drivers/scsi/emcctd/Makefile > > @@ -0,0 +1 @@ > > +obj-$(CONFIG_SCSI_EMCCTD) += emcctd.o > > diff --git a/drivers/scsi/emcctd/README > b/drivers/scsi/emcctd/README > > new file mode 100644 > > index 0000000..496ce7f > > --- /dev/null > > +++ b/drivers/scsi/emcctd/README > > @@ -0,0 +1,10 @@ > > +EMCCTD : Client driver implementation for EMC-Symmetrix > GuestOS emulated > > +Cut-Through Device. The Cut-Through Device PCI emulation is > implemented for > > +GuestOS environments in the HyperMax OS. GuestOS > environments allows loading of > > +any x86 compliant operating systems like Linux/FreeBSD etc. > > + > > +The client driver is a SCSI HBA implementation which interfaces with > SCSI > > +midlayer in the north-bound interfaces and connects with the > emulated PCI device > > +on the south side. > > + > > +The PCI vendor ID:product ID for emulated Cut-Through Device is > 0x1120:0x1B00. > > diff --git a/drivers/scsi/emcctd/emc_ctd_interface.h > b/drivers/scsi/emcctd/emc_ctd_interface.h > > new file mode 100644 > > index 0000000..58a0276 > > --- /dev/null > > +++ b/drivers/scsi/emcctd/emc_ctd_interface.h > > @@ -0,0 +1,386 @@ > > +/* > > + * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem. > > + * > > + * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA. > > + * > > + * Authors: > > + * fredette, matt <matt.fredette@xxxxxxx> > > + * Pirotte, Serge <serge.pirotte@xxxxxxx> > > + * Singh Animesh <Animesh.Singh@xxxxxxx> > > + * Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > > + * > > + * This program is free software; you can redistribute it and/or > > + * modify it under the terms of the GNU General Public License > > + * as published by the Free Software Foundation; either version 2 > > + * of the License, or (at your option) any later version. > > + * > > + * This program is distributed in the hope that it will be useful, > > + * but WITHOUT ANY WARRANTY; without even the implied > warranty of > > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See > the > > + * GNU General Public License for more details. > > + * > > + */ > > + > > +#ifndef _EMC_CTD_INTERFACE_H > > +#define _EMC_CTD_INTERFACE_H > > +/* > > + * > > + * DESCRIPTION: Data structures used by the Cut-through > subsystem. > > + * > > + * NOTES: Changes to any of these structures will mean that any > Clients that > > + * depend on them will also need to be modified. Since many of > those > > + * Clients are not part of the Enginuity build process, this will > almost > > + * certainly require new versions of Guest OS CTD Client code to > be > > + * released. > > + * > > + * Before modifying any data structures in this file please discuss > the > > + * change with the maintainers. > > + * > > + */ > > + > > + > > +/* macros: */ > > + > > +/* the PCI vendor ID for all devices: */ > > +#define EMC_CTD_PCI_VENDOR > (0x1120) > > + > > +/* the PCI product ID for all version 1.x devices: */ > > +#define EMC_CTD_V010_PCI_PRODUCT > (0x1b00) > > + > > +/* the PCI revision ID for the first version 1.0 device: */ > > +#define EMC_CTD_V010_PCI_REVISION (1) > > + > > +/* the 64-bit BAR pair for the transmit and receive rings: */ > > +#define EMC_CTD_V010_BAR_RINGS > (0) > > + > > +/* the 64-bit BAR pair for the fast registers: */ > > +#define EMC_CTD_V010_BAR_FREGS > (2) > > + > > +/* the 32-bit BAR for the slow registers: */ > > +#define EMC_CTD_V010_BAR_SREGS > (4) > > + > > +/* the maximum number of immediate SGL entries: */ > > +#define EMC_CTD_V010_SGL_IMMEDIATE_MAX > (7) > > + > > +/* the CTD v010 whats: */ > > +#define EMC_CTD_V010_WHAT_DETECT (0) > > +#define EMC_CTD_V010_WHAT_SCSI_COMMAND > (1) > > +#define EMC_CTD_V010_WHAT_SCSI_PHASE > (2) > > +#define EMC_CTD_V010_WHAT_SCSI_RESPONSE > (3) > > + > > +/* the CTD v010 detect flags. all undefined flags must be zero: */ > > + > > +/* if this is set, the name is a SCSI target: */ > > +#define EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET > (1 << 0) > > + > > +/* if this is set, the name is a SCSI initiator: */ > > +#define EMC_CTD_V010_DETECT_FLAG_SCSI_INITIATOR > (1 << 1) > > + > > +/* the CTD v010 SCSI command flags. all undefined flags must be > zero: */ > > + > > +/* when the guest receives a SCSI command message, this flag is > undefined. > > + * > > + * If this is set, at the beginning of any data phase the target is the > data > > + * source. if this is clear, at the beginning of any data phase the > target is > > + * the data sink. > > + */ > > +#define EMC_CTD_V010_SCSI_COMMAND_FLAG_SOURCE > (1 << 0) > > + > > +/* when the guest receives a SCSI command message, this flag is > undefined. > > + * > > + * if this is set, the first SGL entry in the message points to an > extended SGL, > > + * and the remaining SGL entries in the message are undefined. if > this is > > + * clear, the SGL entries in the message are used: > > + */ > > +#define EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL > (1 << 1) > > + > > +/* the CTD v010 SCSI response flags. all undefined flags must be > zero: */ > > + > > +/* if this is set, the SCSI command failed. if this is clear, the > > + * command succeeded: > > + */ > > +#define EMC_CTD_V010_SCSI_RESPONSE_FLAG_FAILED > (1 << 0) > > + > > +/* if this is set, any extra information is sense data. if this is clear, > any > > + * extra information is 64-bit timestamps: > > + */ > > +#define EMC_CTD_V010_SCSI_RESPONSE_FLAG_SENSE > (1 << 1) > > + > > +/* the CTD v010 SCSI phase flags. all undefined flags must be zero: > */ > > + > > +/* when the guest receives a SCSI phase message, this flag is > undefined. > > + * > > + * if this is set, at this point in the data phase the message receiver > is the > > + * data source. if this is clear, at this point in the data phase the > message > > + * receiver is the data sink: > > + */ > > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_SOURCE > (1 << 0) > > + > > +/* when the guest receives a SCSI phase message, this flag is > undefined. > > + * > > + * if this is set, the first SGL entry in the message points to an > extended SGL, > > + * and the remaining SGL entries in the message are undefined. if > this is > > + * clear, the SGL entries in the message are used: > > + */ > > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_ESGL (1 << 1) > > + > > +/* if this is set, the message receiver is the target. if this is clear, the > > + * message receiver is the initiator: > > + */ > > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET > (1 << 2) > > + > > +/* if this is set, the SCSI command is aborted: */ > > +#define EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT > (1 << 3) > > + > > +/* the size of the log of errored transmit messages: */ > > +#define EMC_CTD_V010_LOG_ERROR_TX_SIZE > (4) > > + > > +/* errors: */ > > + > > +/* no error: */ > > +#define EMC_CTD_V010_ERROR_NULL > (0) > > + > > +/* the guest tried to transmit a message on a disconnected channel: > */ > > +#define EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED > (1) > > + > > +/* the guest tried to transmit a message with a bad what: */ > > +#define EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT > (2) > > + > > +/* the guest tried to transmit a message with a reserved field set to > > + * the wrong value: > > + */ > > +#define EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED > (3) > > + > > +/* the guest tried to transmit an out-of-order message: */ > > +#define EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER > (4) > > + > > +/* the guest tried to transmit a message to an endpoint whose type > > + * doesn't support it: > > + */ > > +#define EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE > (5) > > + > > +/* the guest tried to transmit a message with an unknown message > > + * receiver's opaque value: > > + */ > > +#define EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN > (6) > > + > > +/* types: */ > > + > > +/* a CTD v010 scatter/gather list entry: */ > > +struct emc_ctd_v010_sgl { > > + > > + /* the physical address of the buffer: */ > > + emc_ctd_uint32_t emc_ctd_v010_sgl_paddr_0_31; > > + emc_ctd_uint32_t emc_ctd_v010_sgl_paddr_32_63; > > + > > + /* the size of the buffer: */ > > + emc_ctd_uint32_t emc_ctd_v010_sgl_size; > > +}; > > + > > +/* a CTD v010 header: */ > > +struct emc_ctd_v010_header { > > + > > + /* the other address: */ > > + emc_ctd_uint16_t emc_ctd_v010_header_address; > > + > > + /* the minor version: */ > > + emc_ctd_uint8_t emc_ctd_v010_header_minor; > > + > > + /* the what: */ > > + emc_ctd_uint8_t emc_ctd_v010_header_what; > > +}; > > + > > +/* a CTD v010 name: */ > > +struct emc_ctd_v010_name { > > + > > + /* the name: */ > > + emc_ctd_uint8_t emc_ctd_v010_name_bytes[8]; > > +}; > > + > > +/* a CTD v010 detect message: */ > > +struct emc_ctd_v010_detect { > > + > > + /* the header: */ > > + struct emc_ctd_v010_header emc_ctd_v010_detect_header; > > + > > + /* the flags: */ > > + emc_ctd_uint32_t emc_ctd_v010_detect_flags; > > + > > + /* the name: */ > > + struct emc_ctd_v010_name emc_ctd_v010_detect_name; > > + > > + /* the key: */ > > + emc_ctd_uint64_t emc_ctd_v010_detect_key; > > +}; > > + > > +/* a CTD v010 SCSI command message: */ > > +struct emc_ctd_v010_scsi_command { > > + > > + /* the header: */ > > + struct emc_ctd_v010_header > emc_ctd_v010_scsi_command_header; > > + > > + /* the flags: */ > > + emc_ctd_uint32_t emc_ctd_v010_scsi_command_flags; > > + > > + /* the initiator's opaque value: */ > > + emc_ctd_uint64_t emc_ctd_v010_scsi_command_opaque; > > + > > + /* the SCSI LUN: */ > > + emc_ctd_uint8_t emc_ctd_v010_scsi_command_lun[8]; > > + > > + /* the SCSI CDB: */ > > + emc_ctd_uint8_t emc_ctd_v010_scsi_command_cdb[16]; > > + > > + /* the data size: */ > > + emc_ctd_uint32_t emc_ctd_v010_scsi_command_data_size; > > + > > + union { > > + > > + /* any SGL entries: */ > > + /* when received by the guest, these are undefined: > */ > > + struct emc_ctd_v010_sgl > > + > emc_ctd_v010_scsi_command_sgl[EMC_CTD_V010_SGL_IMM > EDIATE_MAX]; > > + > > + } emc_ctd_v010_scsi_command_u; > > +}; > > + > > +/* a CTD v010 SCSI response message: */ > > +struct emc_ctd_v010_scsi_response { > > + > > + /* the header: */ > > + struct emc_ctd_v010_header > emc_ctd_v010_scsi_response_header; > > + > > + /* the flags: */ > > + emc_ctd_uint16_t emc_ctd_v010_scsi_response_flags; > > + > > + /* the extra information size: */ > > + emc_ctd_uint8_t emc_ctd_v010_scsi_response_extra_size; > > + > > + /* the SCSI status: */ > > + emc_ctd_uint8_t emc_ctd_v010_scsi_response_status; > > + > > + /* the initiator's opaque value: */ > > + emc_ctd_uint64_t emc_ctd_v010_scsi_response_opaque; > > + > > + /* the data size: */ > > + emc_ctd_uint32_t emc_ctd_v010_scsi_response_data_size; > > + > > + union { > > + > > + /* any extra information: */ > > + emc_ctd_uint8_t > emc_ctd_v010_scsi_response_extra[108]; > > + > > + } emc_ctd_v010_scsi_response_u; > > +}; > > + > > +/* a CTD v010 SCSI phase message: */ > > +struct emc_ctd_v010_scsi_phase { > > + > > + /* the header: */ > > + struct emc_ctd_v010_header > emc_ctd_v010_scsi_phase_header; > > + > > + /* the flags: */ > > + emc_ctd_uint32_t emc_ctd_v010_scsi_phase_flags; > > + > > + /* the message receiver's opaque value: */ > > + emc_ctd_uint64_t emc_ctd_v010_scsi_phase_opaque_rx; > > + > > + /* the message transmitter's opaque value: */ > > + emc_ctd_uint64_t emc_ctd_v010_scsi_phase_opaque_tx; > > + > > + union { > > + > > + /* any SGL entries: */ > > + /* when received by the guest, these are undefined: > */ > > + struct emc_ctd_v010_sgl > > + > emc_ctd_v010_scsi_phase_sgl[EMC_CTD_V010_SGL_IMMEDI > ATE_MAX]; > > + > > + } emc_ctd_v010_scsi_phase_u; > > +}; > > + > > +/* a CTD v010 message: */ > > +union emc_ctd_v010_message { > > + > > + /* the header: */ > > + struct emc_ctd_v010_header > emc_ctd_v010_message_header; > > + > > + /* a detect message: */ > > + struct emc_ctd_v010_detect emc_ctd_v010_message_detect; > > + > > + /* a SCSI command message: */ > > + struct emc_ctd_v010_scsi_command > emc_ctd_v010_message_scsi_command; > > + > > + /* a SCSI response message: */ > > + struct emc_ctd_v010_scsi_response > emc_ctd_v010_message_scsi_response; > > + > > + /* a SCSI phase message: */ > > + struct emc_ctd_v010_scsi_phase > emc_ctd_v010_message_scsi_phase; > > + > > + /* padding: */ > > + emc_ctd_uint8_t emc_ctd_v010_message_padding[128]; > > +}; > > + > > +/* the fast registers: */ > > +struct emc_ctd_v010_fregs { > > + > > + /* the transmit ring producer index (TPI): */ > > + volatile emc_ctd_uint32_t > emc_ctd_v010_fregs_tx_index_producer; > > + > > + /* the error flag: */ > > + volatile emc_ctd_uint32_t emc_ctd_v010_fregs_error_flag; > > + > > + /* errors 1..14: */ > > + volatile emc_ctd_uint32_t > emc_ctd_v010_fregs_errors_1_14[14]; > > + > > + /* the transmit ring consumer index (TCI): */ > > + volatile emc_ctd_uint32_t > emc_ctd_v010_fregs_tx_index_consumer; > > + > > + /* the device name: */ > > + struct emc_ctd_v010_name > emc_ctd_v010_fregs_device_name; > > + > > + /* padding: */ > > + emc_ctd_uint8_t > > + emc_ctd_v010_fregs_pad_07f[64 - > > + (sizeof(emc_ctd_uint32_t) + > > + sizeof(struct emc_ctd_v010_name))]; > > + > > + /* the receive ring producer index (RPI): */ > > + volatile emc_ctd_uint32_t > emc_ctd_v010_fregs_rx_index_producer; > > + > > + /* the interrupt throttle, in units of nanoseconds. zero > disables > > + * the throttle: > > + */ > > + emc_ctd_uint32_t > emc_ctd_v010_fregs_interrupt_throttle_nsecs; > > + > > + /* padding: */ > > + emc_ctd_uint8_t > > + emc_ctd_v010_fregs_pad_0bf[64 - > > + (sizeof(emc_ctd_uint32_t) + > sizeof(emc_ctd_uint32_t))]; > > + > > + /* the receive ring consumer index (RCI): */ > > + volatile emc_ctd_uint32_t > emc_ctd_v010_fregs_rx_index_consumer; > > + > > + /* padding: */ > > + emc_ctd_uint8_t > > + emc_ctd_v010_fregs_pad_0ff[64 - > > + (sizeof(emc_ctd_uint32_t) + > > + (sizeof(emc_ctd_uint32_t) * > > + > EMC_CTD_V010_LOG_ERROR_TX_SIZE))]; > > + > > + /* the errors for the log of errored transmit messages: */ > > + volatile emc_ctd_uint32_t > > + > emc_ctd_v010_fregs_log_error_tx_error[EMC_CTD_V010_LO > G_ERROR_TX_SIZE]; > > + > > + /* the log of errored transmit messages: */ > > + union emc_ctd_v010_message > > + > emc_ctd_v010_fregs_log_error_tx_message[EMC_CTD_V010 > _LOG_ERROR_TX_SIZE]; > > +}; > > + > > +/* the slow registers: */ > > +struct emc_ctd_v010_sregs { > > + > > + /* the reset register: */ > > + emc_ctd_uint32_t emc_ctd_v010_sregs_reset; > > +}; > > + > > +#endif /* _EMC_CTD_INTERFACE_H */ > > diff --git a/drivers/scsi/emcctd/emcctd.c > b/drivers/scsi/emcctd/emcctd.c > > new file mode 100644 > > index 0000000..aa6dd0e > > --- /dev/null > > +++ b/drivers/scsi/emcctd/emcctd.c > > @@ -0,0 +1,2840 @@ > > +/* > > + * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem. > > + * > > + * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA. > > + * > > + * Authors: > > + * fredette, matt <matt.fredette@xxxxxxx> > > + * Pirotte, Serge <serge.pirotte@xxxxxxx> > > + * Singh Animesh <Animesh.Singh@xxxxxxx> > > + * Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > > + * > > + * This program is free software; you can redistribute it and/or > > + * modify it under the terms of the GNU General Public License > > + * as published by the Free Software Foundation; either version 2 > > + * of the License, or (at your option) any later version. > > + * > > + * This program is distributed in the hope that it will be useful, > > + * but WITHOUT ANY WARRANTY; without even the implied > warranty of > > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See > the > > + * GNU General Public License for more details. > > + * > > + */ > > + > > +#include <linux/init.h> > > +#include <linux/kernel.h> > > +#include <linux/kthread.h> > > +#include <linux/slab.h> > > +#include <linux/list.h> > > +#include <linux/module.h> > > +#include <linux/spinlock.h> > > +#include <linux/pci.h> > > +#include <linux/interrupt.h> > > +#include <linux/proc_fs.h> > > +#include <linux/version.h> > > +#include <linux/delay.h> > > +#include <linux/blkdev.h> > > +#include <linux/atomic.h> > > + > > +#include <linux/cache.h> > > + > > +#include <linux/seq_file.h> > > + > > +#include <scsi/scsi.h> > > +#include <scsi/scsi_cmnd.h> > > +#include <scsi/scsi_device.h> > > +#include <scsi/scsi_host.h> > > +#include <scsi/scsi_transport.h> > > +#include <scsi/scsi_dbg.h> > > + > > +#define emc_ctd_uint8_t u8 > > +#define emc_ctd_uint16_t u16 > > +#define emc_ctd_uint32_t u32 > > +#define emc_ctd_uint64_t u64 > > + > > +#include "emc_ctd_interface.h" > > + > > +#include "emcctd.h" > > + > > +#include <scsi/scsi_dbg.h> > > + > > +/* nomenclature for versioning > > + * MAJOR:MINOR:SUBVERSION:PATCH > > + */ > > + > > +#define EMCCTD_MODULE_VERSION "2.0.0.24" > > + > > +MODULE_LICENSE("GPL"); > > +MODULE_AUTHOR("EMC"); > > +MODULE_DESCRIPTION("EMC CTD V1 - Build 18-Jan-2016"); > > +MODULE_VERSION(EMCCTD_MODULE_VERSION); > > + > > +int ctd_debug; > > +module_param_named(ctd_debug, ctd_debug, int, S_IRUGO | > S_IWUSR); > > +MODULE_PARM_DESC(ctd_debug, "Enable driver debug > messages(0=off, 1=on)"); > > + > > +static int emcctd_max_luns = EMCCTD_MAX_LUN; > > +module_param_named(max_luns, emcctd_max_luns, int, > S_IRUGO | S_IWUSR); > > +MODULE_PARM_DESC(max_luns, > > + "Specify the maximum number of LUN's per > host(default=16384)"); > > + > > +static int emcctd_cmd_per_lun = EMCCTD_CMD_PER_LUN; > > +module_param_named(cmd_per_lun, emcctd_cmd_per_lun, int, > S_IRUGO | S_IWUSR); > > +MODULE_PARM_DESC(cmd_per_lun, > > + "Specify the maximum commands per > lun(default=64)"); > > + > > +/* > > + * routines to handle scsi cmd buffer > > + */ > > +static int ctd_xmit_command(struct scsi_cmnd *cmnd, > > + struct ctd_pci_private > *ctd_private); > > +static void ctd_handle_scsi_response( > > + struct emc_ctd_v010_scsi_response *io_response, > > + struct ctd_pci_private *ctd_private); > > + > > +static int ctd_hw_enqueue_request(union > emc_ctd_v010_message *ctd_request, > > + struct ctd_pci_private > *ctd_private); > > +static int ctd_hw_dequeue_response(union > emc_ctd_v010_message *ctd_response, > > + struct ctd_pci_private > *ctd_private); > > +static int ctd_scsi_response_sanity_check( > > + struct ctd_request_private *request_private, > > + struct ctd_pci_private *ctd_private); > > + > > +/* > > + * routines to handle memory and queue management between > client and server > > + */ > > +static int ctd_hw_execute_command(struct scsi_cmnd *cmnd, > > + struct ctd_pci_private *ctd_private); > > +static int ctd_initiator_translate_sgl(struct scsi_cmnd *cmnd, > > + struct emc_ctd_v010_scsi_command > *ctd_request, > > + struct ctd_pci_private *ctd_private); > > +static int ctd_initiator_translate_request(struct scsi_cmnd *cmnd, > > + struct emc_ctd_v010_scsi_command > *ctd_request, > > + struct ctd_pci_private *ctd_private); > > +static void ctd_initiator_translate_lun(struct scsi_cmnd *cmnd, > > + struct emc_ctd_v010_scsi_command > *ctd_request); > > + > > +static struct ctd_request_private *ctd_acquire_request( > > + struct ctd_pci_private *ctd_private); > > +static void ctd_release_request(struct ctd_request_private > *ctd_request, > > + struct ctd_pci_private > *ctd_private); > > +static void ctd_release_io_pool(struct ctd_pci_private > *ctd_private); > > +static int ctd_alloc_io_pool(struct ctd_pci_private *ctd_private, > > + unsigned int > pool_size); > > +static void ctd_check_error_condition(struct pci_dev *pci_dev); > > +static int ctd_init_event_thread(struct ctd_pci_private > *ctd_private); > > +static void ctd_destroy_event_thread(struct ctd_pci_private > *ctd_private); > > +/* > > + * routines related to tasklet functionality > > + */ > > +static void ctd_check_response_queue(unsigned long > instance_addr); > > + > > +static int ctd_ITnexus_handler(struct ctd_pci_private *ctd_private); > > +/* > > + * routines related to internal representation of scsi targets > > + */ > > +static int ctd_target_alloc(struct scsi_target *starget); > > +static void ctd_target_destroy(struct scsi_target *starget); > > +/* > > + * routines registered with the linux scsi midlayer > > + */ > > +static int ctd_queuecommand_lck(struct scsi_cmnd *cmnd, > > + void (*done)(struct scsi_cmnd *)); > > +static DEF_SCSI_QCMD(ctd_queuecommand) > > +static int ctd_slave_configure(struct scsi_device *sdp); > > +static int ctd_slave_alloc(struct scsi_device *sdev); > > +static void ctd_slave_destroy(struct scsi_device *sdev); > > +static int ctd_abort_handler(struct scsi_cmnd *cmd); > > + > > +static enum blk_eh_timer_return ctd_timeout_handler(struct > scsi_cmnd *scmd); > > + > > + > > +/* > > + * routines related to initialization of the pseudo hardware > > + */ > > +static void ctd_init_scsi_host_private(struct Scsi_Host *shost, > > + struct pci_dev > *pci_dev); > > + > > +static int ctd_scsi_layer_init(struct pci_dev *pci_dev); > > +static int ctd_scsi_layer_cleanup(struct pci_dev *pci_dev); > > + > > +#ifdef CONFIG_PM > > +static int ctd_pci_suspend(struct pci_dev *pci_dev, pm_message_t > state); > > +static int ctd_pci_resume(struct pci_dev *pci_dev); > > +#endif > > + > > +static void ctd_pci_remove(struct pci_dev *pci_dev); > > +static int ctd_request_msi(struct pci_dev *pci_dev); > > +static int ctd_pci_probe(struct pci_dev *pci_dev, > > + const struct pci_device_id *id); > > +static void ctd_clear_io_queue(struct ctd_pci_private > *ctd_private); > > + > > +#if !defined(__VMKLNX__) > > +static int ctd_proc_init(struct pci_dev *pci_dev); > > +static void ctd_proc_remove(struct pci_dev *pci_dev); > > + > > +static int ctd_proc_open(struct inode *inode, struct file *file); > > +#endif > > + > > +static int ctd_post_event(union emc_ctd_v010_message *io_msg, > > + struct ctd_pci_private *ctd_private); > > +static int ctd_event_handler(void *ctd_thread_args); > > + > > +unsigned int lun_discovery_complete; > > +wait_queue_head_t lun_discovery_event_barrier; > > + > > +static const struct pci_device_id ctd_pci_id_table[] = { > > + { EMC_CTD_PCI_VENDOR, EMC_CTD_V010_PCI_PRODUCT, > > + PCI_ANY_ID, PCI_ANY_ID, 0, 0, 0 }, > > + { 0 }, > > +}; > > + > > +MODULE_DEVICE_TABLE(pci, ctd_pci_id_table); > > + > > +static struct pci_driver ctd_pci_driver = { > > + .name = "emcctd", > > + .id_table = ctd_pci_id_table, > > + .probe = ctd_pci_probe, > > + .remove = ctd_pci_remove, > > +#ifdef CONFIG_PM > > + .suspend = ctd_pci_suspend, > > + .resume = ctd_pci_resume, > > +#endif > > +}; > > + > > +static struct scsi_host_template scsi_ctd_template = { > > + .name = DRV_NAME, > > + .proc_name = DRV_NAME, > > + .queuecommand = ctd_queuecommand, > > + > > + .eh_timed_out = ctd_timeout_handler, > > + > > + .slave_alloc = ctd_slave_alloc, > > + .slave_configure = ctd_slave_configure, > > + .slave_destroy = ctd_slave_destroy, > > + .eh_abort_handler = ctd_abort_handler, > > + .target_alloc = ctd_target_alloc, > > + .target_destroy = ctd_target_destroy, > > + .can_queue = EMCCTD_CMD_PER_LUN, > > + .this_id = EMCCTD_THIS_ID, > > + .sg_tablesize = SG_ALL, > > + .max_sectors = SCSI_DEFAULT_MAX_SECTORS, > > + .cmd_per_lun = EMCCTD_CMD_PER_LUN, > > + .use_clustering = DISABLE_CLUSTERING, > > + .module = THIS_MODULE, > > +}; > > + > > + > > +#if !defined(__VMKLNX__) > > +static struct proc_dir_entry *ctd_proc_directory; > > + > > +static int > > +ctd_proc_show(struct seq_file *m, void *v) > > +{ > > + int i; > > + struct ctd_pci_private *ctd_private = > > + (struct ctd_pci_private *)m->private; > > + > > + seq_printf(m, > > + "number interrupts: %ld\n" > > + "requests queued: %ld\n" > > + "responses received: %ld\n" > > + "pending IO count: %ld\n" > > + "Abort Sent: %ld\n" > > + "Abort received: %ld\n" > > + "What received: %ld\n" > > + "What sent: %ld\n" > > + "free IO entries : %ld\n" > > + "CTD WWN: %x.%x.%x.%x.%x.%x.%x.%x\n", > > + ctd_private->hw_stats.interrupts.counter, > > + ctd_private->hw_stats.requests_sent.counter, > > + ctd_private- > >hw_stats.responses_received.counter, > > + ctd_private- > >hw_stats.active_io_count.counter, > > + ctd_private->hw_stats.abort_sent.counter, > > + ctd_private- > >hw_stats.abort_received.counter, > > + ctd_private->hw_stats.what_in.counter, > > + ctd_private->hw_stats.what_out.counter, > > + ctd_private- > >hw_stats.free_io_entries.counter, > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[0], > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[1], > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[2], > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[3], > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[4], > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[5], > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[6], > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes[7]); > > + > > +#define MAX_ENTRIES_IN_LINE 10 > > + seq_puts(m, "\nIO Latency (in tsc) for last 200 IOs:\n"); > > + > > + for (i = 0; i < CTD_MAX_IO_STATS; i++) { > > + if (0 == (i % MAX_ENTRIES_IN_LINE)) > > + seq_puts(m, "\n"); > > + > > + seq_printf(m, "%lld \t", ctd_private- > >hw_stats.io_stats[i]); > > + } > > + seq_puts(m, "\n"); > > + > > + return 0; > > + > > +} > > + > > +#endif > > + > > +static void > > +scsi_translate_sam_code(struct scsi_cmnd *cmnd, unsigned char > ScsiStatus) > > +{ > > + unsigned char host_status; > > + unsigned char driver_status; > > + > > + host_status = DID_OK; > > + driver_status = DRIVER_OK; > > + cmnd->result |= ScsiStatus & 0xff; > > + > > + switch (ScsiStatus) { > > + > > + case SAM_STAT_GOOD: > > + case SAM_STAT_CONDITION_MET: > > + case SAM_STAT_INTERMEDIATE_CONDITION_MET: > > + case SAM_STAT_INTERMEDIATE: > > + break; > > + case SAM_STAT_CHECK_CONDITION: > > + case SAM_STAT_RESERVATION_CONFLICT: > > + case SAM_STAT_ACA_ACTIVE: > > + driver_status = DRIVER_SENSE; > > + break; > > + case SAM_STAT_TASK_SET_FULL: > > + case SAM_STAT_BUSY: > > + driver_status = DRIVER_BUSY; > > + host_status = DID_REQUEUE; > > + break; > > + case SAM_STAT_TASK_ABORTED: > > + driver_status = DRIVER_ERROR; > > + host_status = DID_ABORT; > > + break; > > + case SAM_STAT_COMMAND_TERMINATED: > > + default: > > + ctd_dprintk_crit( > > + "cmnd = %p [ channel:%d id:%d lun:%lld] > INVALID SAM = %x\n", > > + cmnd, cmnd->device->channel, cmnd- > >device->id, > > + cmnd->device->lun, ScsiStatus); > > + > > + driver_status = DRIVER_INVALID; > > + host_status = DID_ABORT; > > + break; > > + } > > + set_driver_byte(cmnd, driver_status); > > + set_host_byte(cmnd, host_status); > > +} > > + > > +static void > > +scsi_free_ctd_request_private(struct ctd_request_private > *request_private, > > + struct ctd_pci_private *ctd_private) > > +{ > > + > > + if (request_private->cdb_page) { > > + __free_pages(request_private->cdb_page, > > + request_private->cdb_page_order); > > + } > > + if (request_private->sgllist_page) { > > + __free_pages(request_private->sgllist_page, > > + request_private->sgllist_page_order); > > + } > > + > > + ctd_release_request(request_private, ctd_private); > > +} > > + > > +static int > > +ctd_handle_disconnect(struct emc_ctd_v010_detect *io_detect, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int i; > > + int error; > > + struct ctd_target_info *ctd_target; > > + struct ctd_host_info *ctd_host; > > + > > + error = -ENODEV; > > + > > + ctd_dprintk_crit("\n"); > > + > > + ctd_host = ctd_private->host_private; > > + > > + /* Current implementation only handles the disconnect of > > + * the target and not initiators > > + */ > > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > > + ctd_target = &ctd_host->target[i]; > > + > > + /* detect header address is used to uniquely identify > the target > > + * for which the disconnect event has been posted by > the server > > + */ > > + > > + if (ctd_target- > >ctd_detect.emc_ctd_detect_header_address == > > + io_detect->emc_ctd_detect_header_address) > { > > + /* check the current link status of the target */ > > + if (ctd_target- > >ctd_detect.emc_ctd_v010_detect_flags) { > > + ctd_dprintk_crit("\n"); > > + > > + ctd_target- > >ctd_detect.emc_ctd_v010_detect_flags = > > + io_detect- > >emc_ctd_v010_detect_flags; > > + > > + > ctd_check_error_condition(ctd_private->pci_dev); > > + } else { > > + ctd_dprintk_crit( > > + "target %llx already in > disconnect state\n", > > + (emc_ctd_uint64_t)ctd_target- > >ctd_detect_name_bytes[0]); > > + } > > + error = 0; > > + break; > > + } > > + } > > + > > + if (error) > > + ctd_dprintk_crit("Error\n"); > > + > > + return error; > > +} > > + > > +static int > > +ctd_handle_target_addition(struct emc_ctd_v010_detect > *io_detect, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int i; > > + int error; > > + struct ctd_target_info *ctd_target; > > + struct ctd_host_info *ctd_host; > > + > > + error = -ENOMEM; > > + ctd_host = ctd_private->host_private; > > + > > + ctd_dprintk("header addr -> %x key -> %llx\n", > > + io_detect->emc_ctd_detect_header_address, > > + io_detect->emc_ctd_v010_detect_key); > > + > > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > > + ctd_target = &ctd_host->target[i]; > > + > > + /* detect header address is used to uniquely identify > the target > > + * for which the connect event has been posted by the > server > > + * check if this particular target is already recorded with > the > > + * client check if the recorded target is in the correct > state > > + * and if not found record this particular target in the > list of > > + * the detected targets > > + */ > > + > > + if (ctd_target->ctd_detect.emc_ctd_v010_detect_key > == > > + io_detect->emc_ctd_v010_detect_key) { > > + error = 0; > > + ctd_dprintk("\n"); > > + > > + scsi_target_unblock(&ctd_target->starget- > >dev, > > + SDEV_RUNNING); > > + ctd_target- > >ctd_detect.emc_ctd_v010_detect_flags = > > + io_detect- > >emc_ctd_v010_detect_flags; > > + break; > > + } > > + > > + /* End of list for the recorded targets in the client, so > the > > + * reported target is a new target being reported by > the server > > + * thus needs to be added to the list > > + */ > > + if (ctd_target- > >ctd_detect.emc_ctd_v010_detect_flags == 0x0) { > > + error = 0; > > + ctd_dprintk("\n"); > > + ctd_target->ctd_detect = *io_detect; > > + break; > > + } > > + } > > + if (error) > > + ctd_dprintk_crit("Error\n"); > > + > > + return error; > > +} > > + > > +static int > > +ctd_handle_source_addition(struct emc_ctd_v010_detect > *io_detect, > > + struct ctd_pci_private *ctd_private) > > +{ > > + ctd_dprintk("functionality not supported\n"); > > + return -ENODEV; > > +} > > + > > +static int > > +ctd_handle_detect(struct emc_ctd_v010_detect *io_detect, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + > > + /* > > + * We post the detect event in the event queue and return. > While > > + * ctd_event_thread actually consumes the requests in the > event queue. > > + * This is done to serialize the consecutive detect requests > (disconnect > > + * followed by connect). This mechanism handles the situation > when > > + * multiple detect comes in a quick succession. Also, there is a > > + * seperate thread and its queue for each adapter, so detect > requests > > + * for different adapters shall be handled seperately. > > + */ > > + error = ctd_post_event((union emc_ctd_v010_message > *)io_detect, > > + > ctd_private); > > + > > + if (!error) > > + atomic_long_inc(&ctd_private->hw_stats.what_in); > > + > > + return error; > > + > > +} > > + > > +static void > > +ctd_handle_scsi_command(struct emc_ctd_v010_scsi_command > *io_command, > > + struct ctd_pci_private > *ctd_private) > > +{ > > + ctd_dprintk_crit("unsupported\n"); > > +} > > + > > +static void > > +ctd_handle_scsi_phase(struct emc_ctd_v010_scsi_phase > *io_phase, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + unsigned long flags; > > + struct scsi_cmnd *cmnd; > > + struct ctd_request_private *request_private; > > + > > + /* check the phase flag to confirm we have received correct > phase msg */ > > + if (io_phase->emc_ctd_v010_scsi_phase_flags & > > + > EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET) { > > + ctd_dprintk_crit("SCSI_PHASE_FLAG_TARGET not > supported\n"); > > + return; > > + } > > + > > + if (!(io_phase->emc_ctd_v010_scsi_phase_flags & > > + > EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT)) { > > + ctd_dprintk_crit("scsi_phase_flags %x invalid\n", > > + io_phase->emc_ctd_v010_scsi_phase_flags); > > + return; > > + } > > + > > + request_private = (struct ctd_request_private *) > > + ((uintptr_t)(io_phase- > >emc_ctd_v010_scsi_phase_opaque_rx)); > > + > > + error = ctd_scsi_response_sanity_check(request_private, > ctd_private); > > + > > + if (error) > > + return; > > + > > + cmnd = request_private->cmnd; > > + > > + ctd_dprintk_crit("SCSI_PHASE_FLAG_ABORT cmnd-> %p > request -> %p\n", > > + cmnd, request_private); > > + > > + atomic_long_inc(&ctd_private->hw_stats.abort_received); > > + > > + switch (request_private->io_state) { > > + > > + case CTD_IO_REQUEST_QUEUED: > > + case CTD_IO_REQUEST_REQUEUED: > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, > flags); > > + > > + list_del(&request_private->list); > > + atomic_long_dec(&ctd_private- > >hw_stats.active_io_count); > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, > flags); > > + break; > > + > > + case CTD_IO_REQUEST_ABORTED: > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, > flags); > > + > > + list_del(&request_private->list); > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, > flags); > > + > > + case CTD_IO_REQUEST_REPLY_AWAITED: > > + scsi_free_ctd_request_private(request_private, > ctd_private); > > + return; > > + > > + case CTD_IO_REQUEST_FREE: > > + case CTD_IO_REQUEST_INVALID: > > + default: > > + ctd_dprintk_crit("opaque @ %p in unknown state > %x\n", > > + request_private, request_private->io_state); > > + return; > > + } > > + > > + cmnd->host_scribble = NULL; > > + request_private->cmnd = NULL; > > + request_private->io_state = CTD_IO_REQUEST_COMPLETED; > > + scsi_free_ctd_request_private(request_private, ctd_private); > > + > > + /* error propagation to the SCSI midlayer */ > > + scsi_translate_sam_code(cmnd, SAM_STAT_TASK_ABORTED); > > + scsi_set_resid(cmnd, scsi_bufflen(cmnd)); > > + cmnd->scsi_done(cmnd); > > +} > > + > > +static void > > +ctd_handle_response(union emc_ctd_v010_message > *io_message, > > + struct ctd_pci_private *ctd_private) > > +{ > > + struct emc_ctd_v010_header *msg_header; > > + > > + msg_header = &io_message- > >emc_ctd_v010_message_header; > > + > > + switch (msg_header->emc_ctd_v010_header_what) { > > + > > + case EMC_CTD_V010_WHAT_DETECT: > > + ctd_handle_detect((struct emc_ctd_v010_detect > *)io_message, > > + ctd_private); > > + break; > > + case EMC_CTD_V010_WHAT_SCSI_COMMAND: > > + ctd_handle_scsi_command( > > + (struct emc_ctd_v010_scsi_command > *)io_message, > > + ctd_private); > > + break; > > + case EMC_CTD_V010_WHAT_SCSI_PHASE: > > + ctd_handle_scsi_phase( > > + (struct emc_ctd_v010_scsi_phase *) > io_message, > > + ctd_private); > > + break; > > + case EMC_CTD_V010_WHAT_SCSI_RESPONSE: > > + ctd_handle_scsi_response( > > + (struct emc_ctd_v010_scsi_response > *)io_message, > > + ctd_private); > > + break; > > + default: > > + ctd_dprintk_crit("unknown what -> %x ctd_private -> > %p", > > + msg_header->emc_ctd_v010_header_what, > ctd_private); > > + } > > +} > > + > > + > > +static int > > +ctd_scsi_response_sanity_check(struct ctd_request_private > *request_private, > > + struct ctd_pci_private > *ctd_private) > > +{ > > + int rc; > > + unsigned long flags; > > + struct scsi_cmnd *cmnd; > > + struct ctd_request_private *request, *request_next; > > + int outstanding_io_count = 0; > > + > > + rc = -EFAULT; > > + > > + /* check if the opaque is within the valid range */ > > + if (!((request_private >= ctd_private->io_map) && > > + (request_private < ctd_private->io_map_end))) { > > + ctd_dprintk_crit( > > + "ERROR request_private -> %p in invalid > range\n", > > + request_private); > > + goto ctd_scsi_response_sanity_check_complete; > > + } > > + > > + if (request_private) { > > + cmnd = request_private->cmnd; > > + if (cmnd) { > > + /* check if the back pointer is valid before we > declare > > + * request sane > > + */ > > + if ((request_private == (struct > ctd_request_private > > + *)cmnd->host_scribble)) { > > + rc = 0; > > + goto > ctd_scsi_response_sanity_check_complete; > > + } > > + } > > + /* check if the request has already been completed to > the SCSI > > + * Midlayer > > + */ > > + else if ((request_private->io_state == > > + CTD_IO_REQUEST_ABORTED) || > > + (request_private->io_state == > > + > CTD_IO_REQUEST_REPLY_AWAITED)) { > > + rc = 0; > > + goto > ctd_scsi_response_sanity_check_complete; > > + } > > + > > You don't do any cleanup work at > ctd_scsi_response_sanity_check_complete. You > could just reutrn 0 here as well. [MS>] Just avoiding multiple exit points from the function > > > + > > + ctd_dprintk_crit( > > + "ERROR cmnd -> %p mismatched > request_private -> %p host_scribble -> %p requests send -> %ld > responses received -> %ld state -> %s\n", > > + cmnd, request_private, > > + (cmnd ? cmnd->host_scribble : 0x0), > > + ctd_private- > >hw_stats.requests_sent.counter, > > + ctd_private- > >hw_stats.responses_received.counter, > > + (request_private->io_state == > > + CTD_IO_REQUEST_QUEUED ? > > + > "CTD_IO_REQUEST_QUEUED" : > > + (request_private->io_state == > > + CTD_IO_REQUEST_REQUEUED > ? > > + > "CTD_IO_REQUEST_REQUEUED" : > > + (request_private->io_state == > > + CTD_IO_REQUEST_ABORTED ? > > + > "CTD_IO_REQUEST_ABORTED" : > > + (request_private->io_state == > > + CTD_IO_REQUEST_FREE ? > > + > "CTD_IO_REQUEST_FREE" : > > + (request_private->io_state == > > + CTD_IO_REQUEST_INVALID ? > > + > "CTD_IO_REQUEST_INVALID" : > > + (request_private->io_state == > > + > CTD_IO_REQUEST_COMPLETED ? > > + > "CTD_IO_REQUEST_COMPLETED" : > > + (request_private->io_state == > > + > CTD_IO_REQUEST_REPLY_AWAITED ? > > + > "CTD_IO_REQUEST_REPLY_AWAITED" : > > + "UNKNOWN")))))))); > > + > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, > flags); > > + > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private->aborted_io_list, > list) { > > + if (request == request_private) { > > + ctd_dprintk_crit( > > + "request_private -> %p in > aborted_io_list\n", > > + request_private); > > + break; > > + } > > + } > > + > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private->queued_io_list, > list) { > > + if (request == request_private) { > > + ctd_dprintk_crit( > > + "request_private -> %p in > queued_io_list\n", > > + request_private); > > + } > > + outstanding_io_count++; > > + } > > + ctd_dprintk_crit("outstanding_io_count = %d\n", > > + outstanding_io_count); > > + > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private- > >requeued_io_list, list) { > > + if (request == request_private) { > > + ctd_dprintk_crit( > > + "request_private -> %p in > requeued_io_list\n", > > + request_private); > > + break; > > + } > > + } > > + > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private->io_pool, list) { > > + if (request == request_private) { > > + ctd_dprintk_crit( > > + "request_private -> %p in free > io_pool\n", > > + request_private); > > + break; > > + } > > + } > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, > flags); > > + } else { > > + ctd_dprintk_crit("ERROR request_private -> NULL\n"); > > + } > > + > > +ctd_scsi_response_sanity_check_complete: > > + return rc; > > +} > > + > > +static void > > +ctd_handle_scsi_response(struct emc_ctd_v010_scsi_response > *io_response, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + unsigned long flags; > > + struct scsi_cmnd *cmnd; > > + struct ctd_request_private *request_private; > > + unsigned long long io_stats; > > + > > + request_private = (struct ctd_request_private *) > > + ((uintptr_t)(io_response- > >emc_ctd_v010_scsi_response_opaque)); > > + > > + error = ctd_scsi_response_sanity_check(request_private, > ctd_private); > > + > > + if (error) > > + return; > > + > > + io_stats = ctd_read_tsc(); > > + > > + io_stats -= request_private->io_start_time; > > + > > + ctd_private->hw_stats.io_stats[ctd_private- > >hw_stats.io_stats_index++] = > > + io_stats; > > + > > + if (ctd_private->hw_stats.io_stats_index == > CTD_MAX_IO_STATS) > > + ctd_private->hw_stats.io_stats_index = 0; > > + /* > > + * check if the IO context is in the aborted IO list > > + */ > > + > > + /* Increment the responses_received stats */ > > + atomic_long_inc(&ctd_private- > >hw_stats.responses_received); > > + > > + switch (request_private->io_state) { > > + > > + /* > > + * The state of the request is important > > + * CTD_IO_REQUEST_QUEUED : cmnd is still alive and valid in > kernel > > + * CTD_IO_REQUEST_ABORTED : cmnd has already been > handled before > > + * the response from the device > and > > + * only request needs to be > cleaned > > + * up from the abort list > > + * CTD_IO_REQUEST_FREE : represents a state which is > unhandled (unknown) > > + * CTD_IO_REQUEST_REPLY_AWAITED : represents a state > where abort > > + * could not be sent by the timeout > handler > > + * CTD_IO_REQUEST_INVALID : represents a state which is > > + * unhandled (unknown) > > + */ > > + > > + case CTD_IO_REQUEST_QUEUED: > > + case CTD_IO_REQUEST_REQUEUED: > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, > flags); > > + > > + list_del(&request_private->list); > > + request_private->io_state = > CTD_IO_REQUEST_COMPLETED; > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, > flags); > > + break; > > + > > + case CTD_IO_REQUEST_ABORTED: > > + /* > > + * cmnd is already disassociated from the request > private and IO > > + * completed to the SCSI Midlayer by the > timeout|abort handler > > + * delink the request private from the aborted list and > cleanup > > + */ > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, > flags); > > + list_del(&request_private->list); > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, > flags); > > + > > + case CTD_IO_REQUEST_REPLY_AWAITED: > > + /* > > + * return the context back to the io_pool for its reuse > > + */ > > + request_private->io_state = > CTD_IO_REQUEST_COMPLETED; > > + scsi_free_ctd_request_private(request_private, > ctd_private); > > + /* IO already completed to the Midlayer */ > > + return; > > + > > + case CTD_IO_REQUEST_FREE: > > + case CTD_IO_REQUEST_INVALID: > > + default: > > + ctd_dprintk_crit( > > + "opaque @ %p in unknown state %x\n", > > + request_private, request_private->io_state); > > + return; > > + > > + } > > + > > + /* Decrement active_io_count only when request is still > queued. */ > > + atomic_long_dec(&ctd_private->hw_stats.active_io_count); > > + > > + cmnd = request_private->cmnd; > > + > > + cmnd->result = 0; > > + > > + scsi_translate_sam_code(cmnd, > > + io_response- > >emc_ctd_v010_scsi_response_status); > > + > > + scsi_set_resid(cmnd, scsi_bufflen(cmnd) - > > + io_response- > >emc_ctd_v010_scsi_response_data_size); > > + > > + if (io_response->emc_ctd_v010_scsi_response_flags & > > + > EMC_CTD_V010_SCSI_RESPONSE_FLAG_FAILED) { > > + ctd_dprintk_crit( > > + "cmnd = %p CTDCM_FAILED channel:%d id:%d > lun:%lld] status = %x\n", > > + cmnd, cmnd->device->channel, cmnd->device- > >id, > > + cmnd->device->lun, > > + io_response- > >emc_ctd_v010_scsi_response_status); > > + > > + set_host_byte(cmnd, DID_ERROR); > > + if (ctd_debug) { > > + scsi_print_command(cmnd); > > + scsi_print_result(cmnd, NULL, FAILED); > > + scsi_print_sense(cmnd); > > + } > > + } > > + if (io_response->emc_ctd_v010_scsi_response_flags & > > + > EMC_CTD_V010_SCSI_RESPONSE_FLAG_SENSE) { > > + unsigned char *sense_data; > > + unsigned char sense_data_length; > > + > > + sense_data = io_response- > >emc_ctd_scsi_response_extra; > > + sense_data_length = > > + io_response- > >emc_ctd_v010_scsi_response_extra_size; > > + > > + memcpy(cmnd->sense_buffer, sense_data, > sense_data_length); > > + > > + set_driver_byte(cmnd, DRIVER_SENSE); > > + if (ctd_debug) { > > + scsi_print_command(cmnd); > > + scsi_print_result(cmnd, "emcctd sense", > SUCCESS); > > + scsi_print_sense(cmnd); > > + } > > + } > > + > > + if ((io_response->emc_ctd_v010_scsi_response_status == > > + > SAM_STAT_TASK_SET_FULL) || > > + (io_response->emc_ctd_v010_scsi_response_status > == > > + > SAM_STAT_BUSY)) { > > + > > + ctd_dprintk( > > + "QUEUE DEPTH change for channel:%d id:%d > lun:%lld] active io count = %lx\n", > > + cmnd->device->channel, cmnd->device->id, > > + cmnd->device->lun, > > + ctd_private- > >hw_stats.active_io_count.counter); > > + > > + scsi_track_queue_full(cmnd->device, > > + ctd_private- > >hw_stats.active_io_count.counter); > > + } > > + > > + > > + cmnd->host_scribble = NULL; > > + > > + scsi_free_ctd_request_private(request_private, ctd_private); > > + > > + cmnd->scsi_done(cmnd); > > +} > > + > > +static void > > +ctd_scsi_transfer_info(unsigned char *cmd, unsigned long long > *lba, > > + unsigned int *num, unsigned int *ei_lba) > > +{ > > + *ei_lba = 0; > > + > > + switch (*cmd) { > > + > > + case VARIABLE_LENGTH_CMD: > > + *lba = (u64)cmd[19] | (u64)cmd[18] << 8 | > > + (u64)cmd[17] << 16 | (u64)cmd[16] << 24 | > > + (u64)cmd[15] << 32 | (u64)cmd[14] << 40 | > > + (u64)cmd[13] << 48 | (u64)cmd[12] << 56; > > + > > + *ei_lba = (u32)cmd[23] | (u32)cmd[22] << 8 | > > + (u32)cmd[21] << 16 | (u32)cmd[20] << 24; > > + > > + *num = (u32)cmd[31] | (u32)cmd[30] << 8 | > (u32)cmd[29] << 16 | > > + (u32)cmd[28] << 24; > > + break; > > + > > + case WRITE_SAME_16: > > + case WRITE_16: > > + case READ_16: > > + *lba = (u64)cmd[9] | (u64)cmd[8] << 8 | > > + (u64)cmd[7] << 16 | (u64)cmd[6] << 24 | > > + (u64)cmd[5] << 32 | (u64)cmd[4] << 40 | > > + (u64)cmd[3] << 48 | (u64)cmd[2] << 56; > > + > > + *num = (u32)cmd[13] | (u32)cmd[12] << 8 | > (u32)cmd[11] << 16 | > > + (u32)cmd[10] << 24; > > + break; > > + case WRITE_12: > > + case READ_12: > > + *lba = (u32)cmd[5] | (u32)cmd[4] << 8 | (u32)cmd[3] > << 16 | > > + (u32)cmd[2] << 24; > > + > > + *num = (u32)cmd[9] | (u32)cmd[8] << 8 | (u32)cmd[7] > << 16 | > > + (u32)cmd[6] << 24; > > + break; > > + case WRITE_SAME: > > + case WRITE_10: > > + case READ_10: > > + case XDWRITEREAD_10: > > + *lba = (u32)cmd[5] | (u32)cmd[4] << 8 | > (u32)cmd[3] << 16 | > > + (u32)cmd[2] << 24; > > + > > + *num = (u32)cmd[8] | (u32)cmd[7] << 8; > > + break; > > + case WRITE_6: > > + case READ_6: > > + *lba = (u32)cmd[3] | (u32)cmd[2] << 8 | > > + (u32)(cmd[1] & 0x1f) << 16; > > + *num = (cmd[4] == 0) ? 256 : cmd[4]; > > + break; > > + default: > > + break; > > + } > > +} > > + > > + > > +static int > > +ctd_initiator_validate_sgl(struct scsi_cmnd *cmnd, > > + struct emc_ctd_v010_scsi_command > *ctd_request, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int i; > > + int error; > > + struct emc_ctd_v010_sgl *sgl_current; > > + struct emc_ctd_v010_sgl *sgl_extended; > > + emc_ctd_uint64_t sgl_count; > > + unsigned int sgl_buffersize; > > + unsigned long long lba; > > + unsigned int num; > > + unsigned int ei_lba; > > + > > + lba = 0; > > + num = 0; > > + ei_lba = 0; > > + error = -EINVAL; > > + sgl_count = 0; > > + sgl_extended = &ctd_request- > >emc_ctd_scsi_command_sgl[0]; > > + > > + if (ctd_request->emc_ctd_v010_scsi_command_flags & > > + > EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL) { > > + emc_ctd_uint64_t extented_sgl_size; > > + > > + sgl_current = phys_to_virt( > > + ((emc_ctd_uint64_t)sgl_extended- > >emc_ctd_v010_sgl_paddr_32_63 << 32) | > > + ((emc_ctd_uint64_t)sgl_extended- > >emc_ctd_v010_sgl_paddr_0_31 & 0xFFFFFFFF)); > > + > > + extented_sgl_size = sgl_extended- > >emc_ctd_v010_sgl_size; > > + > > + do_div(extented_sgl_size, sizeof(struct > emc_ctd_v010_sgl)); > > + > > + sgl_count = extented_sgl_size; > > + } else { > > + sgl_current = sgl_extended; > > + while (sgl_current < sgl_extended + > > + > EMC_CTD_V010_SGL_IMMEDIATE_MAX) { > > + if (sgl_current->emc_ctd_v010_sgl_size == 0) > > + break; > > + > > + sgl_current++; > > + sgl_count++; > > + } > > + sgl_current = sgl_extended; > > + } > > + > > + if (scsi_sg_count(cmnd) != sgl_count) { > > + scsi_print_command(cmnd); > > + ctd_dprintk_crit( > > + "Mismatched cmnd_sgl_count %d sgl_count = > %lld sgl_size = %d\n", > > + scsi_sg_count(cmnd), > > + sgl_count, sgl_extended- > >emc_ctd_v010_sgl_size); > > + } > > + > > + if (sgl_count) { > > + struct scatterlist *cmnd_sg; > > + int cmnd_sg_count = scsi_sg_count(cmnd); > > + > > + if (cmnd_sg_count > > EMC_CTD_V010_SGL_IMMEDIATE_MAX) { > > + if ((ctd_request- > >emc_ctd_v010_scsi_command_flags & > > + > EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL) == 0) { > > + scsi_print_command(cmnd); > > + ctd_dprintk_crit( > > + "scsi_sg_count = %d but > EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL not set\n", > > + scsi_sg_count(cmnd)); > > + > > + cmnd_sg_count = > EMC_CTD_V010_SGL_IMMEDIATE_MAX; > > + } > > + } > > + > > + scsi_for_each_sg(cmnd, cmnd_sg, cmnd_sg_count, i) { > > + int cmnd_sg_size; > > + emc_ctd_uint64_t cmnd_buffer_pfn; > > + struct page *cmnd_page; > > + emc_ctd_uint32_t cmnd_offset; > > + emc_ctd_uint32_t sgl_size; > > + emc_ctd_uint64_t buffer_pfn; > > + > > + cmnd_page = sg_page(cmnd_sg); > > + cmnd_buffer_pfn = > page_to_phys(cmnd_page); > > + cmnd_sg_size = cmnd_sg->length; > > + cmnd_offset = cmnd_sg->offset; > > + > > + > > + sgl_size = (sgl_current + i)- > >emc_ctd_v010_sgl_size; > > + buffer_pfn = (((emc_ctd_uint64_t) > > + ((sgl_current + i)- > >emc_ctd_v010_sgl_paddr_32_63) << 32) | > > + ((emc_ctd_uint64_t)((sgl_current + i)- > >emc_ctd_v010_sgl_paddr_0_31) & 0xFFFFFFFF)); > > + > > + if ((sgl_size != cmnd_sg_size) || > > + (buffer_pfn != > (cmnd_buffer_pfn + > > + cmnd_offset))) > { > > + scsi_print_command(cmnd); > > + ctd_dprintk_crit( > > + "Mismatch @ i = %d > cmnd_sg_size = %d cmnd_buffer_pfn = %llx sgl_size = %d buffer_pfn > = %llx\n", > > + i, cmnd_sg_size, > cmnd_buffer_pfn, > > + sgl_size, buffer_pfn); > > + } > > + } > > + goto ctd_initiator_validate_sgl_end; > > + } > > + > > + for (i = 0, sgl_buffersize = 0; i < sgl_count; i++, sgl_current++) > > + sgl_buffersize += sgl_current- > >emc_ctd_v010_sgl_size; > > + > > + if (scsi_bufflen(cmnd) && (sgl_buffersize != > scsi_bufflen(cmnd))) { > > + scsi_print_command(cmnd); > > + ctd_dprintk_crit("Mismatched buffer size %d %d\n", > > + scsi_bufflen(cmnd), sgl_buffersize); > > + goto ctd_initiator_validate_sgl_end; > > + } > > + > > + ctd_scsi_transfer_info((unsigned char *)cmnd->cmnd, &lba, > > + &num, &ei_lba); > > + > > + if (num && (sgl_buffersize != (num * 512))) { > > + scsi_print_command(cmnd); > > + ctd_dprintk_crit("Mismatched buffer size %d %d\n", > > + (num * 512), sgl_buffersize); > > + goto ctd_initiator_validate_sgl_end; > > + } > > + error = 0; > > There is no cleanup done at this label so no need to jump here. Reutrn > directly > from the jumping position. [MS>] Yes it could be done... we are just avoiding multiple returns from the function. > > > +ctd_initiator_validate_sgl_end: > > + return error; > > +} > > + > > +static int > > +ctd_initiator_translate_sgl(struct scsi_cmnd *cmnd, > > + struct emc_ctd_v010_scsi_command > *ctd_request, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int i; > > + int size; > > + int error; > > + int sg_count; > > + int rq_count; > > + int embedded_sgl_count; > > + int embedded_sgl_index; > > + struct ctd_request_private *request_private; > > + emc_ctd_uint64_t buffer_pfn; > > + > > + error = FAILED; > > + i = size = 0; > > + embedded_sgl_index = 0; > > + embedded_sgl_count = > EMC_CTD_V010_SGL_IMMEDIATE_MAX; > > + rq_count = 0; > > + > > + request_private = (struct ctd_request_private *)cmnd- > >host_scribble; > > + sg_count = scsi_sg_count(cmnd); > > + > > + if (sg_count > embedded_sgl_count) { > > + struct scatterlist *sg; > > + struct emc_ctd_v010_sgl *sgl_current; > > + struct emc_ctd_v010_sgl *sgl_extended; > > + > > + request_private->sgllist_page_order = > > + get_order((sizeof(struct emc_ctd_v010_sgl) * > sg_count)); > > + request_private->sgllist_page = > > + alloc_pages(GFP_ATOMIC | __GFP_COMP | > __GFP_NOWARN, > > + request_private->sgllist_page_order); > > + > > + if (!request_private->sgllist_page) { > > + ctd_dprintk_crit("alloc_page failure\n"); > > + goto scsi_initiator_translate_sgl_end; > > + } > > + > > + sgl_current = (struct emc_ctd_v010_sgl *) > > + page_address(request_private->sgllist_page); > > + > > + scsi_for_each_sg(cmnd, sg, sg_count, i) { > > +#if defined(__VMKLNX__) > > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > > + sg->cursgel->addr & > 0xFFFFFFFF; > > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 > = > > + sg->cursgel->addr >> > 32; > > + sgl_current->emc_ctd_v010_sgl_size = > sg_dma_len(sg); > > +#else > > + struct page *page; > > + > > + page = sg_page(sg); > > + buffer_pfn = page_to_phys(page); > > + size += sg->length; > > + > > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > > + (buffer_pfn + sg->offset) & > 0xFFFFFFFF; > > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 > = > > + (buffer_pfn + sg->offset) >> > 32; > > + sgl_current->emc_ctd_v010_sgl_size = sg- > >length; > > +#endif > > + sgl_current++; > > + rq_count++; > > + } > > + sgl_extended = &ctd_request- > >emc_ctd_scsi_command_sgl[0]; > > + buffer_pfn = page_to_phys(request_private- > >sgllist_page); > > + > > + sgl_extended->emc_ctd_v010_sgl_paddr_0_31 = > > + buffer_pfn & 0xFFFFFFFF; > > + sgl_extended->emc_ctd_v010_sgl_paddr_32_63 = > buffer_pfn >> 32; > > + sgl_extended->emc_ctd_v010_sgl_size = > > + rq_count * sizeof(struct > emc_ctd_v010_sgl); > > + ctd_request->emc_ctd_v010_scsi_command_flags |= > > + > EMC_CTD_V010_SCSI_COMMAND_FLAG_ESGL; > > + } else { > > + struct scatterlist *sg; > > + struct emc_ctd_v010_sgl *sgl_current; > > + > > + sgl_current = &ctd_request- > >emc_ctd_scsi_command_sgl[0]; > > + > > + scsi_for_each_sg(cmnd, sg, sg_count, i) { > > +#if defined(__VMKLNX__) > > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > > + sg->cursgel->addr & > 0xFFFFFFFF; > > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 > = > > + sg->cursgel->addr >> > 32; > > + sgl_current->emc_ctd_v010_sgl_size = > sg_dma_len(sg); > > +#else > > + struct page *page; > > + > > + size += sg->length; > > + page = sg_page(sg); > > + buffer_pfn = page_to_phys(page); > > + > > + sgl_current->emc_ctd_v010_sgl_paddr_0_31 = > > + (buffer_pfn + sg->offset) & > 0xFFFFFFFF; > > + sgl_current->emc_ctd_v010_sgl_paddr_32_63 > = > > + (buffer_pfn + sg->offset) >> > 32; > > + sgl_current->emc_ctd_v010_sgl_size = sg- > >length; > > +#endif > > + > > + sgl_current++; > > + rq_count++; > > + } > > + } > > + error = SUCCESS; > > + > > + if (ctd_debug) > > + ctd_initiator_validate_sgl(cmnd, ctd_request, > ctd_private); > > + > > +scsi_initiator_translate_sgl_end: > > + return error; > > +} > > + > > + > > +void > > +ctd_initiator_translate_lun(struct scsi_cmnd *cmnd, > > + struct emc_ctd_v010_scsi_command > *ctd_request) > > +{ > > + > > + int i; > > + union ctd_scsi_lun { > > + emc_ctd_uint64_t scsi_lun_64; > > + emc_ctd_uint8_t scsi_lun_8[8]; > > + } ctd_scsi_lun_conversion; > > + > > + ctd_scsi_lun_conversion.scsi_lun_64 = cpu_to_be64(cmnd- > >device->lun); > > + for (i = 0; i < sizeof(ctd_scsi_lun_conversion); i++) { > > + ctd_request->emc_ctd_v010_scsi_command_lun[i] = > > + ctd_scsi_lun_conversion.scsi_lun_8[i]; > > + } > > +} > > + > > +static int > > +ctd_initiator_translate_request(struct scsi_cmnd *cmnd, > > + struct emc_ctd_v010_scsi_command > *ctd_request, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + struct ctd_dev_info *ctd_device; > > + struct ctd_request_private *request_private; > > + int scsi_cdb_size; > > + > > + error = FAILED; > > + request_private = NULL; > > + ctd_device = cmnd->device->hostdata; > > + > > + if (!(ctd_device->ctd_target_detect- > >emc_ctd_v010_detect_flags & > > + > EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) { > > + goto scsi_initiator_translate_request_end; > > + } > > + memset(ctd_request, 0x0, sizeof(struct > emc_ctd_v010_scsi_command)); > > + > > + ctd_request->emc_ctd_scsi_command_header_address = > > + ctd_device->ctd_target_detect- > >emc_ctd_detect_header_address; > > + ctd_request- > >emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_mino > r = > > + > EMCCTD_V010_PROTOCOL_MINOR_VERSION; > > + ctd_request- > >emc_ctd_v010_scsi_command_header.emc_ctd_v010_header_what > = > > + > EMC_CTD_V010_WHAT_SCSI_COMMAND; > > + ctd_request->emc_ctd_v010_scsi_command_flags |= > > + ((cmnd->sc_data_direction == > DMA_FROM_DEVICE) ? > > + > EMC_CTD_V010_SCSI_COMMAND_FLAG_SOURCE : > > + ((cmnd->sc_data_direction == > > + DMA_TO_DEVICE) ? 0 : > 0)); > > + > > + request_private = ctd_acquire_request(ctd_private); > > + > > + if (request_private == NULL) { > > + ctd_dprintk_crit("ctd_acquire_request failure\n"); > > + goto scsi_initiator_translate_request_end; > > + } > > + > > + request_private->cmnd = cmnd; > > + > > + cmnd->host_scribble = (unsigned char *)request_private; > > + > > + ctd_request->emc_ctd_v010_scsi_command_data_size = > scsi_bufflen(cmnd); > > + > > + ctd_request->emc_ctd_v010_scsi_command_opaque = > > + (uintptr_t)request_private; > > + > > + ctd_initiator_translate_lun(cmnd, ctd_request); > > + > > + scsi_cdb_size = cmnd->cmd_len; > > + > > + if (scsi_cdb_size <= > > + sizeof(ctd_request- > >emc_ctd_v010_scsi_command_cdb)) { > > + memcpy(ctd_request- > >emc_ctd_v010_scsi_command_cdb, > > + cmnd->cmnd, > scsi_cdb_size); > > + } else { > > + ctd_dprintk_crit("unsupported scsi cdb size = %d\n", > > + scsi_cdb_size); > > + goto scsi_initiator_translate_request_end; > > + } > > + > > + error = ctd_initiator_translate_sgl(cmnd, ctd_request, > ctd_private); > > + > > +scsi_initiator_translate_request_end: > > + if (error == FAILED) { > > + cmnd->host_scribble = NULL; > > + if (request_private) { > > + > scsi_free_ctd_request_private(request_private, > > + > ctd_private); > > + } > > + } > > + return error; > > +} > > + > > + > > +static int > > +ctd_hw_execute_command(struct scsi_cmnd *cmnd, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + unsigned long flags; > > + struct emc_ctd_v010_scsi_command ctd_request; > > + > > + error = ctd_initiator_translate_request(cmnd, &ctd_request, > > + ctd_private); > > + > > + if (error == SUCCESS) { > > + struct ctd_request_private *request_private; > > + > > + request_private = > > + (struct ctd_request_private *)cmnd- > >host_scribble; > > + > > + /* lock ordering ... io_mgmt_lock followed by isr_lock > > + * ensures that request placed in legitimate queue ... > > + * so that response finds in the correct queue > > + */ > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, > flags); > > + > > + request_private->io_start_time = ctd_read_tsc(); > > + > > + error = ctd_hw_enqueue_request( > > + (union emc_ctd_v010_message > *)&ctd_request, > > + ctd_private); > > + if (error == SUCCESS) { > > + list_add_tail(&request_private->list, > > + &ctd_private- > >queued_io_list); > > + atomic_long_inc(&ctd_private- > >hw_stats.requests_sent); > > + atomic_long_inc(&ctd_private- > >hw_stats.active_io_count); > > + } > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, > flags); > > + > > + if (error != SUCCESS) { > > + ctd_dprintk_crit("ctd_hw_enqueue_request > error\n"); > > + scsi_free_ctd_request_private( > > + (struct ctd_request_private *) > > + (uintptr_t) > > + > (ctd_request.emc_ctd_v010_scsi_command_opaque), > > + ctd_private); > > + } > > + } > > + > > + return error; > > +} > > + > > +static int > > +ctd_hw_enqueue_request(union emc_ctd_v010_message > *ctd_request, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + unsigned long flags; > > + union emc_ctd_v010_message *ctd_request_block; > > + > > + spin_lock_irqsave(&ctd_private->isr_lock, flags); > > + > > + /* > > + * check if any space is available in the array > > + */ > > + ctd_request_block = > > + ((((ctd_private->request_producer_index + 1) % > > + ctd_private->pci_request_queue_size) == > > + ctd_private->request_consumer_index) ? > > + NULL : > > + (ctd_private->pci_request_array + > > + ctd_private- > >request_producer_index)); > > + > > + > > + error = (ctd_request_block ? SUCCESS : FAILED); > > + > > + if (error == SUCCESS) { > > + *ctd_request_block = *ctd_request; > > + > > + ctd_private->request_producer_index = > > + ((ctd_private->request_producer_index + 1) % > > + ctd_private- > >pci_request_queue_size); > > + > > + } > > + > > + spin_unlock_irqrestore(&ctd_private->isr_lock, flags); > > + > > + return error; > > +} > > + > > +static int > > +ctd_hw_dequeue_response(union emc_ctd_v010_message > *ctd_response, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int rc; > > + unsigned long flags; > > + union emc_ctd_v010_message *ctd_response_block; > > + > > + /* protect ourselves from another instance */ > > + spin_lock_irqsave(&ctd_private->isr_lock, flags); > > + > > + ctd_response_block = > > + ((ctd_private->response_consumer_index == > > + ctd_private->response_producer_index) ? > NULL : > > + (ctd_private->pci_response_array + > > + ctd_private- > >response_consumer_index)); > > + > > + rc = ctd_response_block ? SUCCESS : FAILED; > > + > > + if (rc == SUCCESS) { > > + *ctd_response = *ctd_response_block; > > + ctd_private->response_consumer_index = > > + ((ctd_private->response_consumer_index + 1) > % > > + ctd_private->pci_response_queue_size); > > + } else { > > + ctd_check_error_condition(ctd_private->pci_dev); > > + } > > + > > + spin_unlock_irqrestore(&ctd_private->isr_lock, flags); > > + > > + return rc; > > + > > +} > > + > > +static int > > +ctd_xmit_command(struct scsi_cmnd *cmnd, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + > > + cmnd->result = DID_OK; > > + error = ctd_hw_execute_command(cmnd, ctd_private); > > + return error; > > +} > > + > > + > > +static int > > +ctd_queuecommand_lck(struct scsi_cmnd *cmnd, void > (*done)(struct scsi_cmnd *)) > > +{ > > + int error; > > + struct ctd_host_info *ctd_host; > > + struct ctd_pci_private *ctd_private; > > + > > + error = 0; > > + ctd_host = shost_priv(cmnd->device->host); > > + ctd_private = pci_get_drvdata(ctd_host->pci_dev); > > + > > + switch (ctd_private->hw_state) { > > + > > + case CTD_HW_STATE_INITIALIZED: > > + cmnd->scsi_done = done; > > + if (ctd_xmit_command(cmnd, ctd_private) == > SUCCESS) > > + break; > > + > > + case CTD_HW_STATE_DISABLED: > > + cmnd->scsi_done = done; > > + scsi_translate_sam_code(cmnd, > SAM_STAT_TASK_ABORTED); > > + scsi_set_resid(cmnd, scsi_bufflen(cmnd)); > > + cmnd->scsi_done(cmnd); > > + break; > > + default: > > + error = SCSI_MLQUEUE_HOST_BUSY; > > + } > > + > > + return error; > > +} > > + > > +static int > > +ctd_abort_handler(struct scsi_cmnd *cmnd) > > +{ > > + ctd_dprintk_crit("SCSI cmnd -> %p ERROR\n", > > + cmnd); > > + return SUCCESS; > > +} > > + > > + > > +static int > > +ctd_target_alloc(struct scsi_target *starget) > > +{ > > + int error; > > + struct ctd_target_info *ctd_target; > > + struct ctd_host_info *ctd_host; > > + > > + error = -ENODEV; > > + ctd_host = shost_priv(dev_to_shost(&starget->dev)); > > + > > + ctd_dprintk("starget -> %p id -> %x\n", > > + starget, starget->id); > > + > > + ctd_target = &ctd_host->target[starget->id]; > > + > > + /* check for the connection status in the detect flag > > + * also check if the target already registered with the SCSI > midlayer > > + */ > > + if (ctd_target->starget == NULL && > > + (ctd_target- > >ctd_detect.emc_ctd_v010_detect_flags & > > + > EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) { > > + error = 0; > > + ctd_target->starget = starget; > > + starget->hostdata = ctd_target; > > + } else { > > + if (ctd_target->starget != starget) { > > + ctd_dprintk_crit( > > + "failure ctd_target->starget %p and > starget %p dissimilar\n", > > + ctd_target->starget, starget); > > + } else { > > + ctd_dprintk_crit("failure starget %p > unexpected\n", > > + starget); > > + } > > + } > > + > > + return error; > > +} > > + > > +void > > +ctd_target_destroy(struct scsi_target *starget) > > +{ > > + int i; > > + int error; > > + struct ctd_target_info *ctd_target; > > + struct ctd_host_info *ctd_host; > > + > > + error = -ENODEV; > > + ctd_host = shost_priv(dev_to_shost(&starget->dev)); > > + > > + ctd_dprintk_crit("starget @ id = %x\n", starget->id); > > + > > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > > + ctd_target = &ctd_host->target[i]; > > + if (ctd_target->starget == starget) { > > + ctd_target->starget = NULL; > > + error = 0; > > + break; > > + } > > + } > > + > > + if (error) { > > + ctd_dprintk_crit("failure for starget @ id = %x\n", > > + starget->id); > > + } > > +} > > + > > +static int > > +ctd_slave_configure(struct scsi_device *sdevice) > > +{ > > + int error; > > + struct ctd_dev_info *ctd_device; > > + struct ctd_host_info *ctd_host; > > + > > + error = 0; > > + ctd_host = shost_priv(sdevice->host); > > + ctd_device = sdevice->hostdata; > > + > > + /* tune the block layer to generate timout > > + * for the requests queued and reply is awaited > > + */ > > + blk_queue_rq_timeout(sdevice->request_queue, > EMCCTD_REQUEST_TIMEOUT); > > + > > static int ctd_slave_configure(struct scsi_device *sdev) > { > > /* tune the block layer to generate timout > * for the requests queued and reply is awaited > */ > blk_queue_rq_timeout(sdevice->request_queue, > EMCCTD_REQUEST_TIMEOUT); > } [MS>] Yup right. I missed this. > > > + return error; > > +} > > + > > + > > + > > +static int > > +ctd_slave_alloc(struct scsi_device *sdev) > > +{ > > + int error; > > + struct ctd_host_info *ctd_host; > > + struct ctd_target_info *ctd_target; > > + struct ctd_dev_info *ctd_device; > > + > > + error = -ENOMEM; > > + ctd_host = shost_priv(sdev->host); > > + sdev->host->max_cmd_len = EMCCTD_V010_MAX_CDB_SIZE; > > + > > + ctd_device = (struct ctd_dev_info *) > > + kzalloc(sizeof(struct ctd_dev_info), > GFP_ATOMIC); > > Please don't cast the return value of kzalloc() > [MS>] Hmmmm > > + if (ctd_device == NULL) > > return -ENOMEM; > > > + goto ctd_slave_alloc_end; > > + > > + ctd_target = &ctd_host->target[sdev->id]; > > + if (ctd_target->starget) > > + ctd_device->ctd_target_detect = &ctd_target- > >ctd_detect; > > + > > + if (ctd_device->ctd_target_detect) { > > + error = 0; > > + ctd_device->ctd_host = ctd_host; > > + ctd_device->ctd_target = ctd_target; > > + sdev->hostdata = ctd_device; > > + queue_flag_set_unlocked(QUEUE_FLAG_BIDI, sdev- > >request_queue); > > + } else { > > + kfree(ctd_device); > > + error = -ENODEV; > > + } > > + > > +ctd_slave_alloc_end: > > + > > + if (error) { > > + ctd_dprintk_crit("channel = %x id= %x error = %x\n", > > + sdev->channel, sdev->id, error); > > + } > > + return error; > > +} > > + > > +static void > > +ctd_slave_destroy(struct scsi_device *sdev) > > +{ > > + struct ctd_dev_info *dev_info; > > + > > + dev_info = sdev->hostdata; > > + kfree(dev_info); > > +} > > + > > +static enum blk_eh_timer_return > > +ctd_timeout_handler(struct scsi_cmnd *cmd) > > +{ > > + int requeue_error; > > + unsigned long flags; > > + enum blk_eh_timer_return error; > > + struct ctd_host_info *ctd_host; > > + struct ctd_pci_private *ctd_private; > > + struct ctd_request_private *request; > > This variable's name is a bit of confusing (struct request *request ;-). > Maybe ctd_privreq? [MS>] Yes will fix. > > > + unsigned long long tsc_val; > > + > > + requeue_error = FAILED; > > + error = BLK_EH_NOT_HANDLED; > > + ctd_host = shost_priv(cmd->device->host); > > + ctd_private = pci_get_drvdata(ctd_host->pci_dev); > > + > > + > > + request = (struct ctd_request_private *)cmd->host_scribble; > > + > > + tsc_val = ctd_read_tsc(); > > + > > + tsc_val -= request->io_start_time; > > + > > + ctd_dprintk_crit("cmnd -> %p request -> %p, tsc -> %lld\n", > > + cmd, request, tsc_val); > > + > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > > + > > + if (request && request->io_timeout < EMCCTD_MAX_RETRY) { > > The following block is a bit long and therefore it's hard to spot an > eventual > error of handling the io_mgmt_lock. Can't it be factored out in a helper > function? > [MS>] I know its little longer, but actually fits logically together... Will see if it can be factored. Not sure though. > > + struct ctd_dev_info *ctd_device; > > + > > + switch (request->io_state) { > > + > > + /* check if the IO in the queued_io_list */ > > + case CTD_IO_REQUEST_QUEUED: > > + /* check if the IO is already requeued */ > > + case CTD_IO_REQUEST_REQUEUED: > > + /* remove the old IO context from the > requeued_io_list > > + * or queued_io_list > > + */ > > + list_del(&request->list); > > + atomic_long_dec(&ctd_private- > >hw_stats.active_io_count); > > + > > + ctd_device = cmd->device->hostdata; > > + > > + if (!(ctd_device->ctd_target_detect- > >emc_ctd_v010_detect_flags & > > + > EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET)) { > > + ctd_dprintk_crit("device > diconnected\n"); > > + } else { > > + union emc_ctd_v010_message > ctd_message; > > + struct emc_ctd_v010_scsi_phase > *ctd_phase; > > + > > + memset(&ctd_message, 0x0, > > + sizeof(union > emc_ctd_v010_message)); > > + ctd_phase = > > + (struct > emc_ctd_v010_scsi_phase *) & > > + > ctd_message.emc_ctd_v010_message_scsi_phase; > > + > > + /* Need to acertain if this is how an IO > is > > + * aborted by the specification > > + */ > > + > > + /* > > + * OPT-438489 the phase flag needs to > be > > + * initialized with PHASE_FLAG_TARGET > If EMC > > + * CTD V010 SCSI PHASE FLAG TARGET > is set, the > > + * message receiver is the target, > otherwise the > > + * message receiver is the initiator. If > EMC CTD > > + * V010 SCSI PHASE FLAG ABORT is set, > the SCSI > > + * command is aborted. > > + */ > > + ctd_phase- > >emc_ctd_v010_scsi_phase_flags = > > + > EMC_CTD_V010_SCSI_PHASE_FLAG_ABORT | > > + > EMC_CTD_V010_SCSI_PHASE_FLAG_TARGET; > > + > > + ctd_phase- > >emc_ctd_v010_scsi_phase_opaque_tx = > > + > (uintptr_t)request; > > + ctd_phase- > >emc_ctd_v010_scsi_phase_opaque_rx = > > + > (-1); > > + ctd_phase- > >emc_ctd_scsi_phase_header_what = > > + > EMC_CTD_V010_WHAT_SCSI_PHASE; > > + ctd_phase- > >emc_ctd_scsi_phase_header_minor = > > + > EMCCTD_V010_PROTOCOL_MINOR_VERSION; > > + ctd_phase- > >emc_ctd_scsi_phase_header_address = > > + ctd_device- > >ctd_target_detect->emc_ctd_detect_header_address; > > + > > + requeue_error = > > + > ctd_hw_enqueue_request(&ctd_message, > > + > ctd_private); > > + } > > + > > + > > + if (requeue_error != SUCCESS) { > > + /* add the IO context to > requeued_io_list */ > > + /* Client would try to abort the request > in next > > + * timeout (after 20 seconds) > > + */ > > + request->io_state = > CTD_IO_REQUEST_REQUEUED; > > + list_add_tail(&request->list, > > + &ctd_private- > >requeued_io_list); > > + request->io_timeout++; > > + error = BLK_EH_RESET_TIMER; > > + } else { > > + request->cmnd = NULL; > > + cmd->host_scribble = NULL; > > + > > + request->io_state = > CTD_IO_REQUEST_ABORTED; > > + request->purge_lifetime = > > + jiffies + > EMCCTD_OPAQUE_PURGE_WAITTIME; > > + list_add_tail(&request->list, > > + &ctd_private- > >aborted_io_list); > > + atomic_long_inc(&ctd_private- > >hw_stats.abort_sent); > > + > > + /* error propagation to the SCSI > midlayer */ > > + scsi_translate_sam_code(cmd, > > + > SAM_STAT_TASK_ABORTED); > > + scsi_set_resid(cmd, > scsi_bufflen(cmd)); > > + > > + /* indicate no more requeue of this > particular > > + * IO is needed > > + */ > > + error = BLK_EH_HANDLED; > > + } > > + break; > > + > > + default: > > + ctd_dprintk_crit( > > + "request @ %p in unexpected state > %x\n", > > + request, request->io_state); > > + break; > > + } > > + } else if (request) { > > + ctd_dprintk_crit("cmd %p timeout completed io_state > %x\n", > > + cmd, request->io_state); > > + > > + /* remove the old IO context from the > requeued_io_list */ > > + list_del(&request->list); > > + > > + /* > > + * break the context between the cmnd and the > request and > > + * request on the requeue_io_list cannot be reused > until > > + * server replies for the same > > + */ > > + request->cmnd = NULL; > > + cmd->host_scribble = NULL; > > + > > + /* error propagation to the SCSI midlayer */ > > + scsi_translate_sam_code(cmd, > SAM_STAT_TASK_ABORTED); > > + scsi_set_resid(cmd, scsi_bufflen(cmd)); > > + > > + /* we can deallocate the context only once we receive > > + * any reply from the server > > + */ > > + request->io_state = > CTD_IO_REQUEST_REPLY_AWAITED; > > + > > + /* indicate no more requeue of this particular IO is > needed */ > > + error = BLK_EH_HANDLED; > > + } else { > > + ctd_dprintk_crit( > > + "cmnd -> %p request -> NULL error !!!\n", > cmd); > > + } > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > > + > > + return error; > > +} > > + > > + > > +static int > > +ctd_ITnexus_handler(struct ctd_pci_private *ctd_private) > > +{ > > + int i; > > + int error; > > + struct ctd_target_info *ctd_target; > > + struct ctd_host_info *ctd_host; > > + > > + error = 0; > > + > > + ctd_host = (struct ctd_host_info *)ctd_private->host_private; > > + > > + ctd_dprintk_crit("ctd_private -> %p\n", ctd_private); > > + > > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > > + union emc_ctd_v010_message ctd_message; > > + struct emc_ctd_v010_detect *ctd_detect; > > + > > + ctd_target = &ctd_host->target[i]; > > + > > + switch (ctd_target->detect_completed) { > > + > > + case EMCCTD_TARGET_DETECT_NOT_COMPLETED: > > + if (!ctd_target- > >ctd_detect.emc_ctd_v010_detect_flags) > > + break; > > + > > + /* The id defined by the SCSI Midlayer should > match the > > + * index as this routine is indirectly invoked by > the > > + * delayed work mechanism > > + */ > > + > > + ctd_dprintk_crit("ctd_target -> %p index = > %x\n", > > + ctd_target, i); > > + ctd_dprintk_crit("key -> %llx header -> %x\n", > > + ctd_target- > >ctd_detect.emc_ctd_v010_detect_key, > > + ctd_target- > >ctd_detect.emc_ctd_detect_header_address); > > + > > + memset(&ctd_message, 0x0, > > + sizeof(union > emc_ctd_v010_message)); > > + > > + ctd_detect = > &ctd_message.emc_ctd_v010_message_detect; > > + ctd_detect->emc_ctd_v010_detect_flags = > 0x0; > > + ctd_detect->emc_ctd_v010_detect_key = > > + ctd_target- > >ctd_detect.emc_ctd_v010_detect_key; > > + ctd_detect->emc_ctd_detect_header_what = > > + EMC_CTD_V010_WHAT_DETECT; > > + ctd_detect->emc_ctd_detect_header_minor = > > + > EMCCTD_V010_PROTOCOL_MINOR_VERSION; > > + ctd_detect- > >emc_ctd_detect_header_address = > > + ctd_target- > >ctd_detect.emc_ctd_detect_header_address; > > + > > + if (ctd_hw_enqueue_request(&ctd_message, > > + ctd_private) == > SUCCESS) { > > + ctd_dprintk_crit("ctd_target -> %p\n", > > + ctd_target); > > + ctd_target->detect_completed = > > + > EMCCTD_TARGET_DETECT_COMPLETED; > > + > > + atomic_long_inc( > > + &ctd_private- > >hw_stats.what_out); > > + } else { > > + ctd_dprintk_crit( > > + "ctd_target -> %p > ctd_hw_enqueue_request failure\n", > > + ctd_target); > > + error = -EAGAIN; > > + break; > > + } > > + > > + case EMCCTD_TARGET_DETECT_COMPLETED: > > + /* Disconnect case ... we need to remove the > associated > > + * objects from SCSI midlayer > > + */ > > + if (!ctd_target- > >ctd_detect.emc_ctd_v010_detect_flags) { > > + ctd_dprintk_crit("ctd_target -> %p\n", > > + ctd_target); > > + > > + ctd_clear_io_queue(ctd_private); > > + > > + if (ctd_target->starget) { > > + /* the following attempts to clean the > SCSI > > + * midlayer objects > > + */ > > + scsi_target_block( > > + &ctd_target->starget- > >dev); > > + scsi_target_unblock( > > + &ctd_target->starget- > >dev, > > + > SDEV_TRANSPORT_OFFLINE); > > + /* > > + * Target object might still be > active > > + * in case its not reaped > completely (as > > + * with LVM) thus might be > reused when > > + * the link reconnects back > (OPT 443532) > > + */ > > + scsi_remove_target( > > + &ctd_target->starget- > >dev); > > + } else { > > + ctd_dprintk_crit( > > + "starget already > null\n"); > > + } > > + > > + /* declare the link dead and buried */ > > + ctd_target->detect_completed = > > + > EMCCTD_TARGET_DETECT_NOT_COMPLETED; > > + memset(&ctd_target->ctd_detect, > 0x0, > > + sizeof(struct > emc_ctd_v010_detect)); > > + > > + > wake_up(&lun_discovery_event_barrier); > > + } > > + /* Connect case ... need to scan and create the > needed > > + * objects in the SCSI midlayer > > + */ > > + else { > > + ctd_dprintk_crit("ctd_target -> %p\n", > > + ctd_target); > > + scsi_scan_target(&ctd_host->shost- > >shost_gendev, > > + 0, i, SCAN_WILD_CARD, > 1); > > + lun_discovery_complete = 1; > > + > wake_up(&lun_discovery_event_barrier); > > + } > > + break; > > + default: > > + ctd_dprintk_crit("ctd_target -> %p detect > unknown -> %x\n", > > + ctd_target, ctd_target- > >detect_completed); > > + } > > + } > > + > > + return error; > > +} > > + > > +/* This function posts the detect event into adapter specific list */ > > +static int > > +ctd_post_event(union emc_ctd_v010_message *io_msg, > > + struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + struct ctd_event_io_element *event; > > + > > + error = -ENOMEM; > > + > > + event = (struct ctd_event_io_element *) > > + kzalloc(sizeof(struct ctd_event_io_element), > > + > GFP_ATOMIC); > > No cast for kzalloc() return > [MS>] Ok sir. > > + if (event) { > > + error = 0; > > + event->io_msg = *io_msg; > > + spin_lock(&ctd_private->event_io_lock); > > + list_add_tail(&event->list, &ctd_private- > >event_io_list); > > + spin_unlock(&ctd_private->event_io_lock); > > + } else { > > + ctd_dprintk_crit("kzalloc failure\n"); > > Please no out of memory error messages. > > > + } > > + return error; > > +} > > + > > +/* Thread Handler Function. This consumes the events posted into > its queue, and > > + * takes respective action > > + */ > > +static int > > +ctd_event_handler(void *ctd_thread_args) > > +{ > > + struct ctd_event_io_element *event; > > + struct ctd_pci_private *ctd_private; > > + > > + ctd_private = (struct ctd_pci_private *)ctd_thread_args; > > + > > + while (!kthread_should_stop()) { > > + schedule_timeout_interruptible(HZ); > > + > > + event = NULL; > > + > > + spin_lock(&ctd_private->event_io_lock); > > + if (!list_empty(&ctd_private->event_io_list)) { > > + event = list_first_entry(&ctd_private- > >event_io_list, > > + struct ctd_event_io_element, > list); > > + list_del(&event->list); > > + } > > + spin_unlock(&ctd_private->event_io_lock); > > + > > + > > + if (event) { > > + int error; > > + emc_ctd_uint8_t what; > > + union emc_ctd_v010_message *io_msg; > > + struct emc_ctd_v010_detect *io_detect; > > + > > + io_msg = &event->io_msg; > > + what = io_msg- > >emc_ctd_scsi_message_header_what; > > + > > + if (what != EMC_CTD_V010_WHAT_DETECT) { > > + ctd_dprintk_crit("event -> %p what -> > %x\n", > > + event, what); > > + } else { > > + error = -ENODEV; > > + io_detect = > > + &io_msg- > >emc_ctd_v010_message_detect; > > + > > + if (io_detect- > >emc_ctd_v010_detect_flags == > > + 0x0) { > > + error = > ctd_handle_disconnect( > > + io_detect, > ctd_private); > > + } else { > > + if (io_detect- > >emc_ctd_v010_detect_flags & > > + > EMC_CTD_V010_DETECT_FLAG_SCSI_TARGET) { > > + ctd_dprintk( > > + "header addr - > > %x key -> %llx\n", > > + io_detect- > >emc_ctd_detect_header_address, > > + io_detect- > >emc_ctd_v010_detect_key); > > + error = > ctd_handle_target_addition(io_detect, > > + > ctd_private); > > + } > > + if (io_detect- > >emc_ctd_v010_detect_flags & > > + > EMC_CTD_V010_DETECT_FLAG_SCSI_INITIATOR) { > > + ctd_dprintk("\n"); > > + error = > ctd_handle_source_addition(io_detect, > > + > ctd_private); > > + } > > + } > > + if (!error) { > > + int retry = > EMCCTD_DETECT_RETRY_MAX; > > + > > + error = > ctd_ITnexus_handler(ctd_private); > > + /* In case of ITnexus_handler > failure, > > + * pause for 2 seconds before > retrying > > + * the operation again > > + */ > > + while (error && retry) { > > + > schedule_timeout_interruptible(HZ * 2); > > + error = > ctd_ITnexus_handler(ctd_private); > > + retry--; > > + } while (error && retry); > > + > > + } > > + } > > + kfree(event); > > + } > > + } > > + return 0; > > +} > > + > > +static int > > +ctd_init_event_thread(struct ctd_pci_private *ctd_private) > > +{ > > + int error; > > + > > + INIT_LIST_HEAD(&ctd_private->event_io_list); > > + spin_lock_init(&ctd_private->event_io_lock); > > + > > + /* Create the daemon thread to handle detect requests */ > > + ctd_private->ctd_event_thread = > kthread_create(ctd_event_handler, > > + (void *)ctd_private, > "emcctd_event_thread"); > > + error = (!IS_ERR(ctd_private->ctd_event_thread)) ? 0 : - > EBUSY; > > + if (!error) { > > + wake_up_process(ctd_private->ctd_event_thread); > > + } else { > > + ctd_dprintk_crit( > > + "FAILURE, ctd_private -> %p\n", ctd_private); > > + } > > + return error; > > +} > > + > > +static void > > +ctd_destroy_event_thread(struct ctd_pci_private *ctd_private) > > +{ > > + if (ctd_private->ctd_event_thread) > > + kthread_stop(ctd_private->ctd_event_thread); > > +} > > + > > +static void > > +ctd_init_scsi_host_private(struct Scsi_Host *shost, struct pci_dev > *pci_dev) > > +{ > > + int i; > > + struct ctd_host_info *ctd_host_info; > > + struct ctd_pci_private *ctd_private; > > + > > + ctd_private = pci_get_drvdata(pci_dev); > > + > > + ctd_dprintk("ctd_private -> %p\n", ctd_private); > > + > > + ctd_host_info = shost_priv(shost); > > + memset(ctd_host_info, 0x0, sizeof(struct ctd_host_info)); > > + for (i = 0; i < EMCCTD_MAX_ID; i++) { > > + struct ctd_target_info *ctd_target; > > + > > + ctd_target = &ctd_host_info->target[i]; > > + /* nothing to do; > > + * ctd_target->ctd_detect.emc_ctd_v010_detect_flags > already zero > > + */ > > + } > > + > > + > > + ctd_host_info->shost = shost; > > + ctd_host_info->pci_dev = pci_dev; > > + > > + shost->can_queue = ctd_private->pci_request_queue_size; > > + shost->cmd_per_lun = min(emcctd_cmd_per_lun, shost- > >can_queue); > > + shost->max_lun = emcctd_max_luns; > > + shost->max_id = EMCCTD_MAX_ID; > > + > > + ctd_private->host_private = ctd_host_info; > > + > > + ctd_dprintk("scsi_ctd_host = %p\n", ctd_host_info); > > +} > > + > > + > > +static int > > +ctd_scsi_layer_init(struct pci_dev *pci_dev) > > +{ > > + int error; > > + struct Scsi_Host *scsi_ctd_host; > > + struct ctd_pci_private *ctd_private = NULL; > > + > > + ctd_dprintk("pci_dev -> %p\n", pci_dev); > > + > > + scsi_ctd_host = scsi_host_alloc(&scsi_ctd_template, > > + sizeof(struct ctd_host_info)); > > + if (scsi_ctd_host == NULL) { > > + error = -ENOMEM; > > + goto ctd_scsi_layer_init_complete; > return -ENOMEM; > > + } > > + > > + ctd_init_scsi_host_private(scsi_ctd_host, pci_dev); > > + > > + ctd_private = pci_get_drvdata(pci_dev); > > Maybe: > > > + > > + error = ctd_init_event_thread(ctd_private); > > + if (error) > > + goto ctd_scsi_layer_init_complete; > goto out_put_scsi_host; > > + > > + /* > > + * register the HBA to the Linux SCSI stack > > + */ > > + error = scsi_add_host(scsi_ctd_host, &pci_dev->dev); > if (error) > goto out_destroy_thread; > > reutrn 0; > > out_put_scsi_host: > scsi_host_put(scsi_ctd_host); > out_destroy_thread: > ctd_destroy_event_thread(ctd_private); > > return error; [MS>] Hmmm... will recheck this > > + > > +ctd_scsi_layer_init_complete: > > + if (error) { > > + ctd_dprintk_crit("failure, error = %x\n", error); > > + if (scsi_ctd_host != NULL) > > + scsi_host_put(scsi_ctd_host); > > + > > + if (ctd_private) > > + ctd_destroy_event_thread(ctd_private); > > + } > > + return error; > > +} > > + > > +static void > > +ctd_clear_io_queue(struct ctd_pci_private *ctd_private) > > +{ > > + struct list_head iochain; > > + unsigned long flags; > > + struct ctd_request_private *request, *request_next; > > + > > + ctd_dprintk_crit("ctd_private -> %p\n", ctd_private); > > + > > + INIT_LIST_HEAD(&iochain); > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > > + > > + /* post reset need to cleanup the aborted io > > + * as no reply is expected on them > > + */ > > + /* request is still kept as REPLY_AWAITED to > > + * handle any response post connect > > + */ > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private->aborted_io_list, list) { > > + list_del(&request->list); > > + request->io_state = > CTD_IO_REQUEST_REPLY_AWAITED; > > + } > > + > > + /* rifle thru queued and requeued IO list and mark them for > abort, > > + * the completion to the upper layers is handled by the > timeout logic > > + * invoked from the SCSI midlayer > > + */ > > + /* request is still kept as REPLY_AWAITED to handle any > response post > > + * connect > > + */ > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private->queued_io_list, list) { > > + list_del(&request->list); > > + list_add(&request->list, &iochain); > > + request->cmnd->host_scribble = NULL; > > + request->io_state = > CTD_IO_REQUEST_REPLY_AWAITED; > > + > > + /* These requests shall be aborted to upper layer, so > treat them > > + * as abort_sent > > + */ > > + atomic_long_inc(&ctd_private->hw_stats.abort_sent); > > + atomic_long_dec(&ctd_private- > >hw_stats.active_io_count); > > + } > > + > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private->requeued_io_list, list) { > > + list_del(&request->list); > > + list_add(&request->list, &iochain); > > + request->cmnd->host_scribble = NULL; > > + request->io_state = > CTD_IO_REQUEST_REPLY_AWAITED; > > + > > + /* These requests shall be aborted to upper layer, so > treat them > > + * as abort_sent > > + */ > > + atomic_long_inc(&ctd_private->hw_stats.abort_sent); > > + atomic_long_dec(&ctd_private- > >hw_stats.active_io_count); > > + } > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > > + > > + list_for_each_entry_safe(request, request_next, > > + &iochain, list) { > > + struct scsi_cmnd *cmnd; > > + > > + list_del(&request->list); > > + > > + ctd_dprintk_crit( > > + "cmnd -> %p request -> %p > CTD_IO_REQUEST_REPLY_AWAITED\n", > > + request->cmnd, request); > > + > > + cmnd = request->cmnd; > > + request->cmnd = NULL; > > + > > + /* error propagation to the SCSI midlayer */ > > + scsi_translate_sam_code(cmnd, > SAM_STAT_TASK_ABORTED); > > + scsi_set_resid(cmnd, scsi_bufflen(cmnd)); > > + cmnd->scsi_done(cmnd); > > + } > > + ctd_dprintk("ctd_private -> %p\n", ctd_private); > > +} > > + > > +static int > > +ctd_scsi_layer_cleanup(struct pci_dev *pci_dev) > > +{ > > + int error; > > + struct ctd_pci_private *ctd_private; > > + struct ctd_host_info *ctd_host_info; > > + > > + error = 0; > > + > > + ctd_private = pci_get_drvdata(pci_dev); > > + > > + ctd_dprintk("ctd_private pci_dev -> %p %p\n", ctd_private, > pci_dev); > > + > > + ctd_check_response_queue((unsigned long)pci_dev); > > + > > + ctd_clear_io_queue(ctd_private); > > + > > + ctd_destroy_event_thread(ctd_private); > > + > > + flush_scheduled_work(); > > + > > + ctd_host_info = ctd_private->host_private; > > + > > + scsi_remove_host(ctd_host_info->shost); > > + > > + scsi_host_put(ctd_host_info->shost); > > + > > + return error; > > +} > > + > > +#ifdef CONFIG_PM > > +static int > > +ctd_pci_suspend(struct pci_dev *pci_dev, pm_message_t state) > > +{ > > + pci_save_state(pci_dev); > > + pci_set_power_state(pci_dev, PCI_D3hot); > > + return 0; > > +} > > + > > +static int > > +ctd_pci_resume(struct pci_dev *pci_dev) > > +{ > > + pci_restore_state(pci_dev); > > + pci_set_power_state(pci_dev, PCI_D0); > > + return 0; > > +} > > +#endif > > + > > +static void > > +ctd_pci_remove(struct pci_dev *pci_dev) > > +{ > > + struct ctd_pci_private *ctd_private; > > + > > + ctd_dprintk("pic_dev -> %p\n", pci_dev); > > + > > + ctd_private = pci_get_drvdata(pci_dev); > > + > > + ctd_private->hw_state = CTD_HW_STATE_DISABLED; > > + > > + ctd_scsi_layer_cleanup(pci_dev); > > + > > +#if !defined(__VMKLNX__) > > + ctd_proc_remove(pci_dev); > > +#endif > > + free_irq(pci_dev->irq, pci_dev); > > + > > + pci_disable_msi(pci_dev); > > + > > + if (ctd_private->ioaddr_txrx_rings) > > + pci_iounmap(pci_dev, ctd_private- > >ioaddr_txrx_rings); > > + > > + if (ctd_private->ioaddr_fast_registers) > > + pci_iounmap(pci_dev, ctd_private- > >ioaddr_fast_registers); > > + > > + if (ctd_private->ioaddr_slow_registers) > > + pci_iounmap(pci_dev, ctd_private- > >ioaddr_slow_registers); > > + > > + tasklet_kill(&ctd_private->isr_tasklet); > > + > > + ctd_release_io_pool(ctd_private); > > + > > + kfree(ctd_private); > > + > > + pci_release_regions(pci_dev); > > + pci_set_drvdata(pci_dev, NULL); > > + pci_disable_device(pci_dev); > > +} > > + > > +static void > > +ctd_check_error_condition(struct pci_dev *pci_dev) > > +{ > > +#define EMCCTD_MAX_CACHED_ERROR 14 > > + int i, j; > > + int error; > > + union emc_ctd_v010_message message; > > + static emc_ctd_uint32_t > internal_errors_1_14[EMCCTD_MAX_CACHED_ERROR]; > > + struct ctd_pci_private *ctd_private = > pci_get_drvdata(pci_dev); > > + > > + if (ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_error_flag) { > > + for (i = 0; i < EMCCTD_MAX_CACHED_ERROR; i++) { > > + if (internal_errors_1_14[i] != > > + ctd_private->pci_fast_registers- > >emc_ctd_v010_fregs_errors_1_14[i]) { > > + > > + internal_errors_1_14[i] = > > + ctd_private- > >pci_fast_registers->emc_ctd_v010_fregs_errors_1_14[i]; > > + > > + error = i + 1; > > + > > + for (j = 0; j < > EMC_CTD_V010_LOG_ERROR_TX_SIZE; j++) { > > + if (ctd_private- > >pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_error[j] == > > + error) { > > + memcpy(&message, > > + &ctd_private- > >pci_fast_registers->emc_ctd_v010_fregs_log_error_tx_message[j], > > + > sizeof(message)); > > + ctd_dprintk_crit( > > + "header addr - > > %x error -> %s\n", > > + > message.emc_ctd_scsi_message_header_address, > > + (error == > > + > EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED ? > > + > "EMC_CTD_V010_ERROR_TX_CHANNEL_DISCONNECTED" : > > + error > == > > + > EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT ? > > + > "EMC_CTD_V010_ERROR_TX_MESSAGE_WHAT" : > > + error > == > > + > EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED ? > > + > "EMC_CTD_V010_ERROR_TX_MESSAGE_RESERVED" : > > + error > == > > + > EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER ? > > + > "EMC_CTD_V010_ERROR_TX_MESSAGE_ORDER" : > > + error > == > > + > EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE ? > > + > "EMC_CTD_V010_ERROR_TX_ENDPOINT_TYPE" : > > + error > == > > + > EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN ? > > + > "EMC_CTD_V010_ERROR_TX_OPAQUE_RX_UNKNOWN" : > > + > "EMC_CTD_V010_ERROR_NULL")); > > + } > > + } > > + } > > + } > > + } > > +} > > +/* > > + * ctd_check_response_queue > > + * > > + * Bottom half of interrupt handler. > > + */ > > +static void > > +ctd_check_response_queue(unsigned long instance_addr) > > +{ > > + struct pci_dev *pci_dev = (struct pci_dev *)instance_addr; > > + struct ctd_pci_private *ctd_private = > pci_get_drvdata(pci_dev); > > + union emc_ctd_v010_message io_response; > > + > > + > > + /* empty response queue */ > > + while (ctd_hw_dequeue_response(&io_response, > ctd_private) == > > + SUCCESS) { > > + /* handle the response */ > > + ctd_handle_response(&io_response, > ctd_private); > > + } > > +} > > + > > + > > +static irqreturn_t > > +ctd_isr(int irq, void *opaque) > > +{ > > + struct pci_dev *pci_dev = (struct pci_dev *)opaque; > > + struct ctd_pci_private *ctd_private = > pci_get_drvdata(pci_dev); > > + > > + atomic_long_inc(&ctd_private->hw_stats.interrupts); > > + > > + /* schedule work for later */ > > + tasklet_schedule(&ctd_private->isr_tasklet); > > + > > + return IRQ_HANDLED; > > +} > > + > > +static int > > +ctd_request_msi(struct pci_dev *pci_dev) > > +{ > > + int err = -EFAULT; > > + > > + if (pci_dev->irq) { > > + err = pci_enable_msi(pci_dev); > > + if (!err) { > > + err = request_irq(pci_dev->irq, ctd_isr, > > + IRQF_SHARED, > > + pci_name(pci_dev), pci_dev); > > + if (err < 0) { > > + ctd_dprintk_crit("request_irq failure > !!!\n"); > > + pci_disable_msi(pci_dev); > > + err = -EBUSY; > > + } > > + } > > + } > > + return err; > > +} > > + > > +static struct ctd_request_private * > > +ctd_acquire_request(struct ctd_pci_private *ctd_private) > > +{ > > + unsigned long flags; > > + struct ctd_request_private *ctd_request; > > + > > + ctd_request = NULL; > > + > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > > + > > + /* check if any request in the aborted io list can be reused */ > > + if (!list_empty(&ctd_private->aborted_io_list)) { > > + struct ctd_request_private *request, *request_next; > > + > > + list_for_each_entry_safe(request, request_next, > > + &ctd_private->aborted_io_list, > list) { > > + /* aborted_io_list is in chronologically order > thus > > + * failure of time_after() indicates any request > after > > + * this point is not in the kill zone > > + */ > > + if (time_before(jiffies, request- > >purge_lifetime)) > > + break; > > + > > + list_del(&request->list); > > + if (request->cdb_page) { > > + __free_pages(request->cdb_page, > > + request- > >cdb_page_order); > > + } > > + if (request->sgllist_page) { > > + __free_pages(request->sgllist_page, > > + request- > >sgllist_page_order); > > + } > > + memset(request, 0x0, > > + sizeof(struct > ctd_request_private)); > > + list_add(&request->list, &ctd_private- > >io_pool); > > + atomic_long_inc(&ctd_private- > >hw_stats.free_io_entries); > > + } > > + } > > + > > + if (!list_empty(&ctd_private->io_pool)) { > > + ctd_request = list_first_entry(&ctd_private->io_pool, > > + struct ctd_request_private, > list); > > + list_del(&ctd_request->list); > > + ctd_request->io_state = CTD_IO_REQUEST_QUEUED; > > + } > > + > > + if (ctd_request) > > + atomic_long_dec(&ctd_private- > >hw_stats.free_io_entries); > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > > + > > + return ctd_request; > > +} > > + > > +static void > > +ctd_release_request(struct ctd_request_private *ctd_request, > > + struct ctd_pci_private *ctd_private) > > +{ > > + unsigned long flags; > > + > > + spin_lock_irqsave(&ctd_private->io_mgmt_lock, flags); > > + memset(ctd_request, 0x0, sizeof(struct > ctd_request_private)); > > + list_add(&ctd_request->list, &ctd_private->io_pool); > > + > > + atomic_long_inc(&ctd_private->hw_stats.free_io_entries); > > + > > + spin_unlock_irqrestore(&ctd_private->io_mgmt_lock, flags); > > +} > > + > > +static void > > +ctd_release_io_pool(struct ctd_pci_private *ctd_private) > > +{ > > + kfree(ctd_private->io_map); > > +} > > + > > +static int > > +ctd_alloc_io_pool(struct ctd_pci_private *ctd_private, unsigned int > pool_size) > > +{ > > + int i; > > + int error; > > + > > + error = -ENOMEM; > > + > > + INIT_LIST_HEAD(&ctd_private->io_pool); > > + INIT_LIST_HEAD(&ctd_private->queued_io_list); > > + INIT_LIST_HEAD(&ctd_private->aborted_io_list); > > + INIT_LIST_HEAD(&ctd_private->requeued_io_list); > > + > > + spin_lock_init(&ctd_private->io_mgmt_lock); > > + > > + ctd_private->io_map = kcalloc(pool_size, > > + sizeof(struct ctd_request_private), > GFP_KERNEL); > > + > > + /* > > + * in case of allocation failure try with one fourth the size > before > > + * throwing the towel > > + */ > > + > > + if (ctd_private->io_map == NULL) { > > + pool_size = pool_size >> 2; > > + ctd_private->io_map = kcalloc(pool_size, > > + sizeof(struct ctd_request_private), > GFP_KERNEL); > > + if (ctd_private->io_map == NULL) { > > + ctd_dprintk_crit( > > + "io_pool allocation failure for pool_size > -> %d\n", > > + pool_size); > > No OOM messages. [MS>] Just wanted to have a personalized OOM message... to take care of debugging... Will handle it other way round > > > + goto ctd_alloc_io_pool_complete; > > + } > > + } > > + > > + ctd_private->io_map_end = ctd_private->io_map + pool_size; > > + > > + for (i = 0; i < pool_size; i++) { > > + struct ctd_request_private *request_context = > > + ctd_private->io_map + > i; > > + memset(request_context, 0x0, > > + sizeof(struct ctd_request_private)); > > + list_add(&request_context->list, &ctd_private- > >io_pool); > > + } > > + ctd_dprintk_crit( > > + "ctd_private -> %p, pool_size -> %x, io_map -> %p, > io_map_end-> %p\n", > > + ctd_private, pool_size, > > + ctd_private->io_map, ctd_private- > >io_map_end); > > + error = 0; > > + ctd_private->hw_stats.free_io_entries.counter = pool_size; > > +ctd_alloc_io_pool_complete: > > + return error; > > +} > > + > > +static int > > +ctd_pci_probe( > > + struct pci_dev *pci_dev, > > + const struct pci_device_id *id > > +) > > +{ > > + struct ctd_pci_private *ctd_private; > > + int ctd_proc_initialized; > > + int ctd_scsi_initialized; > > + int ctd_regions_initialized; > > + int err; > > + > > + err = -ENODEV; > > + ctd_proc_initialized = FAILED; > > + ctd_scsi_initialized = FAILED; > > + ctd_regions_initialized = FAILED; > > + > > + ctd_private = (struct ctd_pci_private *) > > + kzalloc(sizeof(struct ctd_pci_private), > GFP_ATOMIC); > > + > > + if (ctd_private == NULL) { > > + ctd_dprintk_crit("kzalloc Failure\n"); > > + goto ctd_pci_probe_complete; > > + } > > No cast and no OOM messages. [MS>] Fine. > > > + > > + ctd_private->pci_dev = pci_dev; > > + > > + /* enable the device */ > > + err = pci_enable_device(pci_dev); > > + if (err) { > > + ctd_dprintk_crit("pci_enable_device Failure\n"); > > I'd prefere dev_err() here, as you'll have the information which device > enable > failed. [MS>] ok > > > + goto ctd_pci_probe_complete; > > + } > > + pci_set_master(pci_dev); > > + > > + err = pci_request_regions(pci_dev, "ctd-pci"); > > + if (err) { > > + ctd_dprintk_crit("pci_request_regions Failure\n"); > > Same here. > > > + goto ctd_pci_probe_complete; > > + } > > + > > + ctd_regions_initialized = SUCCESS; > > + > > + ctd_dprintk("ctd_private pci_dev -> %p %p\n", ctd_private, > pci_dev); > > + > > +#define EMC_CTD_TXRX_MSG_SIZE 128 > > + > > + if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_RINGS)) > { > > + ctd_private->ioaddr_txrx_rings = ioremap( > > + pci_resource_start(pci_dev, > > + EMC_CTD_V010_BAR_RINGS), > > + pci_resource_len(pci_dev, > > + EMC_CTD_V010_BAR_RINGS) > > + ); > > + ctd_private->txrx_ringsize = > > + (pci_resource_len(pci_dev, > > + EMC_CTD_V010_BAR_RINGS) >> 1) > > + / EMC_CTD_TXRX_MSG_SIZE; > > + > > + ctd_dprintk_crit( > > + "physical addr = %llx ioaddr_txrx_rings = %p , > ring size = %x\n", > > + pci_resource_start(pci_dev, > > + EMC_CTD_V010_BAR_RINGS), > > + ctd_private->ioaddr_txrx_rings, > > + ctd_private->txrx_ringsize); > > + > > + } > > + if (ctd_private->ioaddr_txrx_rings == NULL) { > > + err = -ENOMEM; > > + ctd_dprintk_crit("ioremap failure\n"); > > + goto ctd_pci_probe_complete; > > + } else { > > + ctd_private->pci_request_array = ctd_private- > >ioaddr_txrx_rings; > > + ctd_private->pci_response_array = > > + ctd_private->ioaddr_txrx_rings + > > + ((pci_resource_len(pci_dev, > > + > EMC_CTD_V010_BAR_RINGS)) >> 1); > > + } > > + > > + > > + if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_FREGS)) > { > > + ctd_private->ioaddr_fast_registers = > > + ioremap(pci_resource_start(pci_dev, > > + EMC_CTD_V010_BAR_FREGS), > > + pci_resource_len(pci_dev, > > + EMC_CTD_V010_BAR_FREGS)); > > + > > + ctd_dprintk_crit( > > + "physical addr = %llx ioaddr_fast_registers = > %p\n", > > + pci_resource_start(pci_dev, > > + EMC_CTD_V010_BAR_FREGS), > > + ctd_private->ioaddr_fast_registers); > > + } > > + if (ctd_private->ioaddr_fast_registers == NULL) { > > + err = -ENOMEM; > > + ctd_dprintk_crit("ioremap failure\n"); > > + goto ctd_pci_probe_complete; > > + } else { > > + ctd_private->pci_fast_registers = > > + ctd_private->ioaddr_fast_registers; > > + } > > + > > + if (pci_resource_start(pci_dev, EMC_CTD_V010_BAR_SREGS)) > { > > + ctd_private->ioaddr_slow_registers = > > + ioremap(pci_resource_start(pci_dev, > > + > EMC_CTD_V010_BAR_SREGS), > > + pci_resource_len(pci_dev, > > + EMC_CTD_V010_BAR_SREGS)); > > + > > + ctd_dprintk_crit( > > + "physical addr = %llx ioaddr_slow_registers = > %p\n", > > + pci_resource_start(pci_dev, > > + EMC_CTD_V010_BAR_SREGS), > > + ctd_private->ioaddr_slow_registers); > > + } > > + if (ctd_private->ioaddr_slow_registers == NULL) { > > + err = -ENOMEM; > > + ctd_dprintk_crit("ioremap failure\n"); > > + goto ctd_pci_probe_complete; > > + } else { > > + ctd_private->pci_slow_registers = > > + ctd_private->ioaddr_slow_registers; > > + } > > + > > + /* reset the device */ > > + ctd_private->pci_device_reset_register = 0XFF; > > + > > + ctd_private->pci_request_queue_size = ctd_private- > >txrx_ringsize; > > + ctd_private->pci_response_queue_size = ctd_private- > >txrx_ringsize; > > + > > + err = ctd_alloc_io_pool(ctd_private, > > + ctd_private->pci_request_queue_size); > > + if (err) { > > + ctd_dprintk_crit("ctd_alloc_io_pool failure\n"); > > + goto ctd_pci_probe_complete; > > + } > > + > > + pci_set_drvdata(pci_dev, ctd_private); > > + > > + spin_lock_init(&ctd_private->isr_lock); > > + > > + /* setup tasklet for scanning response queue */ > > + tasklet_init(&ctd_private->isr_tasklet, > > + ctd_check_response_queue, (unsigned > long)pci_dev); > > + > > + ctd_private->hw_state = CTD_HW_STATE_INITIALIZED; > > + > > + pci_set_master(pci_dev); > > + > > +#if !defined(__VMKLNX__) > > + err = ctd_proc_init(pci_dev); > > + if (err) { > > + ctd_dprintk_crit("ctd_proc_init failure\n"); > > + goto ctd_pci_probe_complete; > > + } > > + ctd_proc_initialized = SUCCESS; > > +#endif > > + err = ctd_scsi_layer_init(pci_dev); > > + if (err) { > > + ctd_dprintk_crit("ctd_scsi_layer_init failure\n"); > > + goto ctd_pci_probe_complete; > > + } > > + ctd_scsi_initialized = SUCCESS; > > + > > + err = ctd_request_msi(pci_dev); > > + if (err) { > > + ctd_dprintk_crit("ctd_request_msi failure\n"); > > + goto ctd_pci_probe_complete; > > + } > > + > > + /* after we reset the device, but before we enabled MSI, > some messages > > + * may have been received. check for them > > + */ > > + tasklet_schedule(&ctd_private->isr_tasklet); > > + > > +ctd_pci_probe_complete: > > + if (err) { > > + if (ctd_private) { > > + tasklet_kill(&ctd_private->isr_tasklet); > > + > > + if (ctd_scsi_initialized == SUCCESS) > > + ctd_scsi_layer_cleanup(pci_dev); > > + > > + if (ctd_proc_initialized == SUCCESS) > > + ctd_proc_remove(pci_dev); > > + > > + if (ctd_private->ioaddr_txrx_rings) > > + pci_iounmap(pci_dev, > > + ctd_private- > >ioaddr_txrx_rings); > > + > > + if (ctd_private->ioaddr_fast_registers) > > + pci_iounmap(pci_dev, > > + ctd_private- > >ioaddr_fast_registers); > > + > > + if (ctd_private->ioaddr_slow_registers) > > + pci_iounmap(pci_dev, > > + ctd_private- > >ioaddr_slow_registers); > > + > > + if (ctd_regions_initialized == SUCCESS) > > + pci_release_regions(pci_dev); > > + > > + ctd_release_io_pool(ctd_private); > > + kfree(ctd_private); > > + } > > + pci_set_drvdata(pci_dev, NULL); > > + pci_disable_device(pci_dev); > > + } > > + return err; > > +} > > + > > +#if !defined(__VMKLNX__) > > +static const struct file_operations ctd_proc_fops = { > > + .open = ctd_proc_open, > > + .read = seq_read, > > + .llseek = seq_lseek, > > + .release = seq_release, > > +}; > > +static int > > +ctd_proc_open(struct inode *inode, struct file *file) > > +{ > > + return single_open(file, ctd_proc_show, PDE_DATA(inode)); > > +} > > + > > +int > > +ctd_proc_init(struct pci_dev *pci_dev) > > +{ > > + int err; > > + static int hw_index; > > + char hw_name[MAX_PROC_FILE_NAMELEN]; > > + struct ctd_pci_private *ctd_private; > > + struct proc_dir_entry *pde; > > + > > + ctd_private = (struct ctd_pci_private *) > pci_get_drvdata(pci_dev); > > + ctd_private->hw_index = hw_index; > > + > > + memset(hw_name, 0x0, sizeof(hw_name)); > > + snprintf(hw_name, sizeof(hw_name), "emcctd_stats_%d", > hw_index++); > > + > > + err = -EPERM; > > + > > + do { > > + if (ctd_proc_directory == NULL) > > + break; > > + pde = proc_create_data(hw_name, S_IFREG | > S_IRUGO | S_IWUSR, > > + ctd_proc_directory, > > + &ctd_proc_fops, > > + ctd_private); > > + > > + if (pde == NULL) { > > + ctd_dprintk_crit( > > + "create_proc_read_entry failure for > %s\n", > > + hw_name); > > + break; > > + } > > + err = 0; > > + } while (0); > > + return err; > > +} > > + > > +void > > +ctd_proc_remove(struct pci_dev *pci_dev) > > +{ > > + int hw_index; > > + char hw_name[MAX_PROC_FILE_NAMELEN]; > > + struct ctd_pci_private *ctd_private; > > + > > + ctd_private = (struct ctd_pci_private *) > pci_get_drvdata(pci_dev); > > + > > + hw_index = ctd_private->hw_index; > > + memset(hw_name, 0x0, sizeof(hw_name)); > > + snprintf(hw_name, sizeof(hw_name), > "emc/emcctd_stats_%d", hw_index); > > + > > + ctd_dprintk("removing %s\n", hw_name); > > + > > + remove_proc_entry(hw_name, NULL); > > + > > +} > > +#endif > > + > > +static int __init ctd_pci_init(void) > > +{ > > + int err; > > + > > + ctd_dprintk_crit("Loading emcctd\n"); > > + init_waitqueue_head(&lun_discovery_event_barrier); > > + > > + ctd_proc_directory = proc_mkdir("emc", NULL); > > + > > + err = pci_register_driver(&ctd_pci_driver); > > + > > + if (err) { > > + remove_proc_entry("emc", NULL); > > + } else { > > + /* wait for 20 seconds or less to allow the luns to > appear > > + * before exiting from insmod > > + */ > > + > wait_event_interruptible_timeout(lun_discovery_event_barri > er, > > + lun_discovery_complete, ((HZ) > * 20)); > > + } > > + > > + return err; > > +} > > + > > +module_init(ctd_pci_init); > > + > > +static void __exit ctd_pci_exit(void) > > +{ > > + pci_unregister_driver(&ctd_pci_driver); > > + remove_proc_entry("emc", NULL); > > +} > > + > > +module_exit(ctd_pci_exit); > > diff --git a/drivers/scsi/emcctd/emcctd.h > b/drivers/scsi/emcctd/emcctd.h > > new file mode 100644 > > index 0000000..3fba931 > > --- /dev/null > > +++ b/drivers/scsi/emcctd/emcctd.h > > @@ -0,0 +1,232 @@ > > +/* > > + * EMCCTD: EMC Cut-Through HBA Driver for SCSI subsystem. > > + * > > + * Copyright (C) 2015 by EMC Corporation, Hopkinton, MA. > > + * > > + * Authors: > > + * fredette, matt <matt.fredette@xxxxxxx> > > + * Pirotte, Serge <serge.pirotte@xxxxxxx> > > + * Singh Animesh <Animesh.Singh@xxxxxxx> > > + * Singhal, Maneesh <Maneesh.Singhal@xxxxxxx> > > + * > > + * This program is free software; you can redistribute it and/or > > + * modify it under the terms of the GNU General Public License > > + * as published by the Free Software Foundation; either version 2 > > + * of the License, or (at your option) any later version. > > + * > > + * This program is distributed in the hope that it will be useful, > > + * but WITHOUT ANY WARRANTY; without even the implied > warranty of > > + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See > the > > + * GNU General Public License for more details. > > + * > > + */ > > + > > +#ifndef _EMCCTD_H_ > > +#define _EMCCTD_H_ > > + > > +#define DRV_NAME "emcctd" > > + > > +#define EMCCTD_V010_PROTOCOL_MINOR_VERSION 0x0 > > + > > +/* please refer to emc_ctd_v010_scsi_command_cdb in > emc_ctd_interface.h */ > > +#define EMCCTD_V010_MAX_CDB_SIZE 16 > > + > > +#define EMCCTD_MAX_LUN 16384 > > +#define EMCCTD_MAX_ID 16 > > +#define EMCCTD_MAX_RETRY 5 > > +#define EMCCTD_CMD_PER_LUN 16 > > +#define EMCCTD_THIS_ID (-1) > > +#define EMCCTD_REQUEST_TIMEOUT (60 * HZ) > > +#define EMCCTD_OPAQUE_PURGE_WAITTIME (10 * HZ) > > + > > +#define EMCCTD_DEVICE_RESET_PAUSE 3 > > +#define EMCCTD_DETECT_RETRY_MAX 3 > > + > > +#define ctd_dprintk(__m_fmt, ...) \ > > +do { \ > > + if (ctd_debug) \ > > + pr_info("%s:%d:"__m_fmt, __func__, __LINE__, > ##__VA_ARGS__); \ > > +} while (0) > > + > > +#define ctd_dprintk_crit(__m_fmt, ...) \ > > + pr_crit("%s:%d:"__m_fmt, __func__, __LINE__, > ##__VA_ARGS__) > > + > > +#define EMCCTD_TARGET_DETECT_COMPLETED 1 > > +#define EMCCTD_TARGET_DETECT_NOT_COMPLETED 0 > > + > > +struct ctd_target_info { > > + unsigned int detect_completed; > > + struct scsi_target *starget; > > + struct emc_ctd_v010_detect ctd_detect; > > +}; > > + > > +#define emc_ctd_detect_header_address \ > > + > emc_ctd_v010_detect_header.emc_ctd_v010_header_addres > s > > +#define emc_ctd_detect_header_minor \ > > + > emc_ctd_v010_detect_header.emc_ctd_v010_header_minor > > +#define emc_ctd_detect_header_what \ > > + > emc_ctd_v010_detect_header.emc_ctd_v010_header_what > > + > > +#define emc_ctd_scsi_command_header_address \ > > + > emc_ctd_v010_scsi_command_header.emc_ctd_v010_header > _address > > +#define emc_ctd_scsi_command_header_minor \ > > + > emc_ctd_v010_scsi_command_header.emc_ctd_v010_header > _minor > > +#define emc_ctd_scsi_command_header_what \ > > + > emc_ctd_v010_scsi_command_header.emc_ctd_v010_header > _what > > + > > +#define emc_ctd_scsi_response_header_address \ > > + > emc_ctd_v010_scsi_response_header.emc_ctd_v010_header > _address > > +#define emc_ctd_scsi_response_header_minor \ > > + > emc_ctd_v010_scsi_response_header.emc_ctd_v010_header > _minor > > +#define emc_ctd_scsi_response_header_what \ > > + > emc_ctd_v010_scsi_response_header.emc_ctd_v010_header > _what > > + > > +#define emc_ctd_scsi_phase_header_address \ > > + > emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_ad > dress > > +#define emc_ctd_scsi_phase_header_minor \ > > + > emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_mi > nor > > +#define emc_ctd_scsi_phase_header_what \ > > + > emc_ctd_v010_scsi_phase_header.emc_ctd_v010_header_w > hat > > + > > +#define emc_ctd_scsi_message_header_address \ > > + > emc_ctd_v010_message_header.emc_ctd_v010_header_addr > ess > > +#define emc_ctd_scsi_message_header_minor \ > > + > emc_ctd_v010_message_header.emc_ctd_v010_header_min > or > > +#define emc_ctd_scsi_message_header_what \ > > + > emc_ctd_v010_message_header.emc_ctd_v010_header_wha > t > > + > > +#define emc_ctd_scsi_command_sgl \ > > + > emc_ctd_v010_scsi_command_u.emc_ctd_v010_scsi_comma > nd_sgl > > +#define emc_ctd_scsi_response_extra \ > > + > emc_ctd_v010_scsi_response_u.emc_ctd_v010_scsi_respons > e_extra > > + > > +#define ctd_detect_name_bytes \ > > + > ctd_detect.emc_ctd_v010_detect_name.emc_ctd_v010_nam > e_bytes > > + > > +struct ctd_host_info { > > + struct Scsi_Host *shost; > > + struct pci_dev *pci_dev; > > + struct ctd_target_info target[EMCCTD_MAX_ID]; > > +}; > > + > > +struct ctd_dev_info { > > + struct ctd_host_info *ctd_host; > > + struct ctd_target_info *ctd_target; > > + struct emc_ctd_v010_detect *ctd_target_detect; > > +}; > > + > > +#define PROC_STAT_SCSI_TS_MAX 10 > > +#define MAX_PROC_FILE_NAMELEN 128 > > +#define CTD_MAX_IO_STATS 200 > > + > > +struct ctd_hw_stats { > > + atomic_long_t interrupts; > > + atomic_long_t requests_sent; > > + atomic_long_t responses_received; > > + atomic_long_t active_io_count; > > + atomic_long_t abort_sent; > > + atomic_long_t abort_received; > > + atomic_long_t what_in; > > + atomic_long_t what_out; > > + atomic_long_t free_io_entries; > > + unsigned long long io_stats[CTD_MAX_IO_STATS]; > > + unsigned int io_stats_index; > > +}; > > + > > +enum ctd_io_request_state { > > + CTD_IO_REQUEST_FREE, > > + CTD_IO_REQUEST_QUEUED, > > + CTD_IO_REQUEST_REQUEUED, > > + CTD_IO_REQUEST_ABORTED, > > + CTD_IO_REQUEST_COMPLETED, > > + CTD_IO_REQUEST_REPLY_AWAITED, > > + CTD_IO_REQUEST_INVALID > > +}; > > + > > +enum ctd_hw_state { > > + CTD_HW_STATE_UNINITIALIZED, > > + CTD_HW_STATE_INITIALIZED, > > + CTD_HW_STATE_UNDER_RESET, > > + CTD_HW_STATE_DISABLED, > > + CTD_HW_STATE_INVALID > > +}; > > + > > +struct ctd_request_private { > > + struct list_head list; > > + enum ctd_io_request_state io_requeue_state; > > + unsigned int io_timeout; > > + enum ctd_io_request_state io_state; > > + struct scsi_cmnd *cmnd; > > + struct page *cdb_page; > > + unsigned int cdb_page_order; > > + struct page *sgllist_page; > > + unsigned int sgllist_page_order; > > + unsigned long purge_lifetime; > > + unsigned long long io_start_time; > > +}; > > + > > +struct ctd_pci_private { > > + struct pci_dev *pci_dev; > > + void *host_private; > > + > > + void __iomem *ioaddr_txrx_rings; > > + void __iomem *ioaddr_fast_registers; > > + void __iomem *ioaddr_slow_registers; > > + > > + emc_ctd_uint32_t txrx_ringsize; > > + emc_ctd_uint32_t pci_request_queue_size; > > + emc_ctd_uint32_t pci_response_queue_size; > > + struct emc_ctd_v010_fregs *pci_fast_registers; > > + struct emc_ctd_v010_sregs *pci_slow_registers; > > + union emc_ctd_v010_message *pci_response_array; > > + union emc_ctd_v010_message *pci_request_array; > > + > > + struct tasklet_struct isr_tasklet; > > + spinlock_t isr_lock; > > + unsigned int hw_index; > > + struct ctd_hw_stats hw_stats; > > + enum ctd_hw_state hw_state; > > + struct list_head queued_io_list; > > + struct list_head aborted_io_list; > > + struct list_head requeued_io_list; > > + struct list_head io_pool; > > + struct ctd_request_private *io_map; > > + struct ctd_request_private *io_map_end; > > + spinlock_t io_mgmt_lock; > > + > > + struct task_struct *ctd_event_thread; > > + struct list_head event_io_list; > > + spinlock_t event_io_lock; > > +}; > > + > > +#define request_producer_index \ > > + pci_fast_registers- > >emc_ctd_v010_fregs_tx_index_producer > > +#define request_consumer_index \ > > + pci_fast_registers- > >emc_ctd_v010_fregs_tx_index_consumer > > +#define response_producer_index \ > > + pci_fast_registers- > >emc_ctd_v010_fregs_rx_index_producer > > +#define response_consumer_index \ > > + pci_fast_registers- > >emc_ctd_v010_fregs_rx_index_consumer > > +#define pci_device_reset_register \ > > + pci_slow_registers->emc_ctd_v010_sregs_reset > > + > > +#define pci_device_name_bytes \ > > + pci_fast_registers- > >emc_ctd_v010_fregs_device_name.emc_ctd_v010_name_bytes > > + > > +struct ctd_event_io_element { > > + struct list_head list; > > + union emc_ctd_v010_message io_msg; > > +}; > > + > > +static inline unsigned long long ctd_read_tsc(void) > > +{ > > + unsigned long long local_tsc; > > + > > + local_tsc = rdtsc(); > > + > > + return local_tsc; > > +} > > + > > +#endif /* _EMCCTD_H_ */ > > + > > +/* vi: set ts=8 sw=8 noet: */ > > -- > > 1.8.5.2 > > > > -- > > To unsubscribe from this list: send the line "unsubscribe linux-scsi" in > > the body of a message to majordomo@xxxxxxxxxxxxxxx > > More majordomo info at http://vger.kernel.org/majordomo- > info.html > > -- > Johannes Thumshirn Storage > jthumshirn@xxxxxxx +49 911 74053 689 > SUSE LINUX GmbH, Maxfeldstr. 5, 90409 Nürnberg > GF: Felix Imendörffer, Jane Smithard, Graham Norton > HRB 21284 (AG Nürnberg) > Key fingerprint = EC38 9CAB C2C4 F25D 8600 D0D0 0393 969D 2D76 0850 -- To unsubscribe from this list: send the line "unsubscribe linux-api" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html