This patch starts the population of the efct target mode driver. The driver is contained in the drivers/scsi/elx/efct subdirectory. This patch creates the efct directory and starts population of the driver by adding SLI-4 configuration parameters, data structures for configuring SLI-4 queues, converting from os to SLI-4 IO requests, and handling async events. Signed-off-by: Ram Vegesna <ram.vegesna@xxxxxxxxxxxx> Signed-off-by: James Smart <jsmart2021@xxxxxxxxx> --- drivers/scsi/elx/efct/efct_hw.h | 852 ++++++++++++++++++++++++++++++++++++++++ 1 file changed, 852 insertions(+) create mode 100644 drivers/scsi/elx/efct/efct_hw.h diff --git a/drivers/scsi/elx/efct/efct_hw.h b/drivers/scsi/elx/efct/efct_hw.h new file mode 100644 index 000000000000..ff6de91923fa --- /dev/null +++ b/drivers/scsi/elx/efct/efct_hw.h @@ -0,0 +1,852 @@ +/* SPDX-License-Identifier: GPL-2.0 */ +/* + * Copyright (C) 2019 Broadcom. All Rights Reserved. The term + * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. + */ + +#ifndef _EFCT_HW_H +#define _EFCT_HW_H + +#include "../libefc_sli/sli4.h" +#include "efct_utils.h" + +/* + * EFCT PCI IDs + */ +#define EFCT_VENDOR_ID 0x10df +/* LightPulse 16Gb x 4 FC (lancer-g6) */ +#define EFCT_DEVICE_ID_LPE31004 0xe307 +#define PCI_PRODUCT_EMULEX_LPE32002 0xe307 +/* LightPulse 32Gb x 4 FC (lancer-g7) */ +#define EFCT_DEVICE_ID_G7 0xf407 + +/*Default RQ entries len used by driver*/ +#define EFCT_HW_RQ_ENTRIES_MIN 512 +#define EFCT_HW_RQ_ENTRIES_DEF 1024 +#define EFCT_HW_RQ_ENTRIES_MAX 4096 + +/*Defines the size of the RQ buffers used for each RQ*/ +#define EFCT_HW_RQ_SIZE_HDR 128 +#define EFCT_HW_RQ_SIZE_PAYLOAD 1024 + +/*Define the maximum number of multi-receive queues*/ +#define EFCT_HW_MAX_MRQS 8 + +/* + * Define count of when to set the WQEC bit in a submitted + * WQE, causing a consummed/released completion to be posted. + */ +#define EFCT_HW_WQEC_SET_COUNT 32 + +/*Send frame timeout in seconds*/ +#define EFCT_HW_SEND_FRAME_TIMEOUT 10 + +/* + * FDT Transfer Hint value, reads greater than this value + * will be segmented to implement fairness. A value of zero disables + * the feature. + */ +#define EFCT_HW_FDT_XFER_HINT 8192 + +#define EFCT_HW_TIMECHECK_ITERATIONS 100 +#define EFCT_HW_MAX_NUM_MQ 1 +#define EFCT_HW_MAX_NUM_RQ 32 +#define EFCT_HW_MAX_NUM_EQ 16 +#define EFCT_HW_MAX_NUM_WQ 32 + +#define OCE_HW_MAX_NUM_MRQ_PAIRS 16 + +#define EFCT_HW_MAX_WQ_CLASS 4 +#define EFCT_HW_MAX_WQ_CPU 128 + +/* + * A CQ will be assinged to each WQ + * (CQ must have 2X entries of the WQ for abort + * processing), plus a separate one for each RQ PAIR and one for MQ + */ +#define EFCT_HW_MAX_NUM_CQ \ + ((EFCT_HW_MAX_NUM_WQ * 2) + 1 + (OCE_HW_MAX_NUM_MRQ_PAIRS * 2)) + +#define EFCT_HW_Q_HASH_SIZE 128 +#define EFCT_HW_RQ_HEADER_SIZE 128 +#define EFCT_HW_RQ_HEADER_INDEX 0 + +/* Options for efct_hw_command() */ +enum { + /* command executes synchronously and busy-waits for completion */ + EFCT_CMD_POLL, + /* command executes asynchronously. Uses callback */ + EFCT_CMD_NOWAIT, +}; + +enum efct_hw_rtn { + EFCT_HW_RTN_SUCCESS = 0, + EFCT_HW_RTN_SUCCESS_SYNC = 1, + EFCT_HW_RTN_ERROR = -1, + EFCT_HW_RTN_NO_RESOURCES = -2, + EFCT_HW_RTN_NO_MEMORY = -3, + EFCT_HW_RTN_IO_NOT_ACTIVE = -4, + EFCT_HW_RTN_IO_ABORT_IN_PROGRESS = -5, + EFCT_HW_RTN_IO_PORT_OWNED_ALREADY_ABORTED = -6, + EFCT_HW_RTN_INVALID_ARG = -7, +}; + +#define EFCT_HW_RTN_IS_ERROR(e) ((e) < 0) + +enum efct_hw_reset { + EFCT_HW_RESET_FUNCTION, + EFCT_HW_RESET_FIRMWARE, + EFCT_HW_RESET_MAX +}; + +enum efct_hw_property { + EFCT_HW_N_IO, + EFCT_HW_N_SGL, + EFCT_HW_MAX_IO, + EFCT_HW_MAX_SGE, + EFCT_HW_MAX_SGL, + EFCT_HW_MAX_NODES, + EFCT_HW_MAX_RQ_ENTRIES, + EFCT_HW_TOPOLOGY, + EFCT_HW_WWN_NODE, + EFCT_HW_WWN_PORT, + EFCT_HW_FW_REV, + EFCT_HW_FW_REV2, + EFCT_HW_IPL, + EFCT_HW_VPD, + EFCT_HW_VPD_LEN, + EFCT_HW_MODE, + EFCT_HW_LINK_SPEED, + EFCT_HW_IF_TYPE, + EFCT_HW_SLI_REV, + EFCT_HW_SLI_FAMILY, + EFCT_HW_RQ_PROCESS_LIMIT, + EFCT_HW_RQ_DEFAULT_BUFFER_SIZE, + EFCT_HW_AUTO_XFER_RDY_CAPABLE, + EFCT_HW_AUTO_XFER_RDY_XRI_CNT, + EFCT_HW_AUTO_XFER_RDY_SIZE, + EFCT_HW_AUTO_XFER_RDY_BLK_SIZE, + EFCT_HW_AUTO_XFER_RDY_T10_ENABLE, + EFCT_HW_AUTO_XFER_RDY_P_TYPE, + EFCT_HW_AUTO_XFER_RDY_REF_TAG_IS_LBA, + EFCT_HW_AUTO_XFER_RDY_APP_TAG_VALID, + EFCT_HW_AUTO_XFER_RDY_APP_TAG_VALUE, + EFCT_HW_DIF_CAPABLE, + EFCT_HW_DIF_SEED, + EFCT_HW_DIF_MODE, + EFCT_HW_DIF_MULTI_SEPARATE, + EFCT_HW_DUMP_MAX_SIZE, + EFCT_HW_DUMP_READY, + EFCT_HW_DUMP_PRESENT, + EFCT_HW_RESET_REQUIRED, + EFCT_HW_FW_ERROR, + EFCT_HW_FW_READY, + EFCT_HW_HIGH_LOGIN_MODE, + EFCT_HW_PREREGISTER_SGL, + EFCT_HW_HW_REV1, + EFCT_HW_HW_REV2, + EFCT_HW_HW_REV3, + EFCT_HW_ETH_LICENSE, + EFCT_HW_LINK_MODULE_TYPE, + EFCT_HW_NUM_CHUTES, + EFCT_HW_WAR_VERSION, + /* enable driver timeouts for target WQEs */ + EFCT_HW_EMULATE_TARGET_WQE_TIMEOUT, + EFCT_HW_LINK_CONFIG_SPEED, + EFCT_HW_CONFIG_TOPOLOGY, + EFCT_HW_BOUNCE, + EFCT_HW_PORTNUM, + EFCT_HW_BIOS_VERSION_STRING, + EFCT_HW_RQ_SELECT_POLICY, + EFCT_HW_SGL_CHAINING_CAPABLE, + EFCT_HW_SGL_CHAINING_ALLOWED, + EFCT_HW_SGL_CHAINING_HOST_ALLOCATED, + EFCT_HW_SEND_FRAME_CAPABLE, + EFCT_HW_RQ_SELECTION_POLICY, + EFCT_HW_RR_QUANTA, + EFCT_HW_FILTER_DEF, + EFCT_HW_MAX_VPORTS, + EFCT_ESOC, +}; + +enum { + EFCT_HW_TOPOLOGY_AUTO, + EFCT_HW_TOPOLOGY_NPORT, + EFCT_HW_TOPOLOGY_LOOP, + EFCT_HW_TOPOLOGY_NONE, + EFCT_HW_TOPOLOGY_MAX +}; + +enum { + EFCT_HW_MODE_INITIATOR, + EFCT_HW_MODE_TARGET, + EFCT_HW_MODE_BOTH, + EFCT_HW_MODE_MAX +}; + +/* pack fw revision values into a single uint64_t */ +#define HW_FWREV(a, b, c, d) (((uint64_t)(a) << 48) | ((uint64_t)(b) << 32) \ + | ((uint64_t)(c) << 16) | ((uint64_t)(d))) + +#define EFCT_FW_VER_STR(a, b, c, d) (#a "." #b "." #c "." #d) + +/* Defines DIF operation modes */ +enum { + EFCT_HW_DIF_MODE_INLINE, + EFCT_HW_DIF_MODE_SEPARATE, +}; + +/* T10 DIF operations */ +enum efct_hw_dif_oper { + EFCT_HW_DIF_OPER_DISABLED, + EFCT_HW_SGE_DIFOP_INNODIFOUTCRC, + EFCT_HW_SGE_DIFOP_INCRCOUTNODIF, + EFCT_HW_SGE_DIFOP_INNODIFOUTCHKSUM, + EFCT_HW_SGE_DIFOP_INCHKSUMOUTNODIF, + EFCT_HW_SGE_DIFOP_INCRCOUTCRC, + EFCT_HW_SGE_DIFOP_INCHKSUMOUTCHKSUM, + EFCT_HW_SGE_DIFOP_INCRCOUTCHKSUM, + EFCT_HW_SGE_DIFOP_INCHKSUMOUTCRC, + EFCT_HW_SGE_DIFOP_INRAWOUTRAW, +}; + +#define EFCT_HW_DIF_OPER_PASS_THRU EFCT_HW_SGE_DIFOP_INCRCOUTCRC +#define EFCT_HW_DIF_OPER_STRIP EFCT_HW_SGE_DIFOP_INCRCOUTNODIF +#define EFCT_HW_DIF_OPER_INSERT EFCT_HW_SGE_DIFOP_INNODIFOUTCRC + +/* T10 DIF block sizes */ +enum efct_hw_dif_blk_size { + EFCT_HW_DIF_BK_SIZE_512, + EFCT_HW_DIF_BK_SIZE_1024, + EFCT_HW_DIF_BK_SIZE_2048, + EFCT_HW_DIF_BK_SIZE_4096, + EFCT_HW_DIF_BK_SIZE_520, + EFCT_HW_DIF_BK_SIZE_4104, + EFCT_HW_DIF_BK_SIZE_NA = 0 +}; + +/* link module types */ +enum { + EFCT_HW_LINK_MODULE_TYPE_1GB = 0x0004, + EFCT_HW_LINK_MODULE_TYPE_2GB = 0x0008, + EFCT_HW_LINK_MODULE_TYPE_4GB = 0x0040, + EFCT_HW_LINK_MODULE_TYPE_8GB = 0x0080, + EFCT_HW_LINK_MODULE_TYPE_10GB = 0x0100, + EFCT_HW_LINK_MODULE_TYPE_16GB = 0x0200, + EFCT_HW_LINK_MODULE_TYPE_32GB = 0x0400, +}; + +/* T10 DIF information passed to the transport */ +struct efct_hw_dif_info { + enum efct_hw_dif_oper dif_oper; + enum efct_hw_dif_blk_size blk_size; + u32 ref_tag_cmp; + u32 ref_tag_repl; + u16 app_tag_cmp; + u16 app_tag_repl; + bool check_ref_tag; + bool check_app_tag; + bool check_guard; + bool auto_incr_ref_tag; + bool repl_app_tag; + bool repl_ref_tag; + bool dif_separate; + + /* If the APP TAG is 0xFFFF, disable REF TAG and CRC field chk */ + bool disable_app_ffff; + + /* if the APP TAG is 0xFFFF and REF TAG is 0xFFFF_FFFF, + * disable checking the received CRC field. + */ + bool disable_app_ref_ffff; + u16 dif_seed; + u8 dif; +}; + +enum efct_hw_io_type { + EFCT_HW_ELS_REQ, + EFCT_HW_ELS_RSP, + EFCT_HW_ELS_RSP_SID, + EFCT_HW_FC_CT, + EFCT_HW_FC_CT_RSP, + EFCT_HW_BLS_ACC, + EFCT_HW_BLS_ACC_SID, + EFCT_HW_BLS_RJT, + EFCT_HW_IO_TARGET_READ, + EFCT_HW_IO_TARGET_WRITE, + EFCT_HW_IO_TARGET_RSP, + EFCT_HW_IO_DNRX_REQUEUE, + EFCT_HW_IO_MAX, +}; + +enum efct_hw_io_state { + EFCT_HW_IO_STATE_FREE, + EFCT_HW_IO_STATE_INUSE, + EFCT_HW_IO_STATE_WAIT_FREE, + EFCT_HW_IO_STATE_WAIT_SEC_HIO, +}; + +struct efct_hw; + +/** + * HW command context. + * Stores the state for the asynchronous commands sent to the hardware. + */ +struct efct_command_ctx { + struct list_head list_entry; + int (*cb)(struct efct_hw *hw, int status, u8 *mqe, void *arg); + void *arg; /* Argument for callback */ + u8 *buf; /* buffer holding command / results */ + void *ctx; /* upper layer context */ +}; + +struct efct_hw_sgl { + uintptr_t addr; + size_t len; +}; + +union efct_hw_io_param_u { + struct { + u16 ox_id; + u16 rx_id; + u8 payload[12]; + } bls; + struct { + u32 s_id; + u16 ox_id; + u16 rx_id; + u8 payload[12]; + } bls_sid; + struct { + u8 r_ctl; + u8 type; + u8 df_ctl; + u8 timeout; + } bcast; + struct { + u16 ox_id; + u8 timeout; + } els; + struct { + u32 s_id; + u16 ox_id; + u8 timeout; + } els_sid; + struct { + u8 r_ctl; + u8 type; + u8 df_ctl; + u8 timeout; + } fc_ct; + struct { + u8 r_ctl; + u8 type; + u8 df_ctl; + u8 timeout; + u16 ox_id; + } fc_ct_rsp; + struct { + u32 offset; + u16 ox_id; + u16 flags; + u8 cs_ctl; + enum efct_hw_dif_oper dif_oper; + enum efct_hw_dif_blk_size blk_size; + u8 timeout; + u32 app_id; + } fcp_tgt; + struct { + struct efc_dma *cmnd; + struct efc_dma *rsp; + enum efct_hw_dif_oper dif_oper; + enum efct_hw_dif_blk_size blk_size; + u32 cmnd_size; + u16 flags; + u8 timeout; + u32 first_burst; + } fcp_ini; +}; + +/* WQ steering mode */ +enum efct_hw_wq_steering { + EFCT_HW_WQ_STEERING_CLASS, + EFCT_HW_WQ_STEERING_REQUEST, + EFCT_HW_WQ_STEERING_CPU, +}; + +/* HW wqe object */ +struct efct_hw_wqe { + struct list_head list_entry; + bool abort_wqe_submit_needed; + bool send_abts; + u32 id; + u32 abort_reqtag; + u8 *wqebuf; +}; + +/** + * HW IO object. + * + * Stores the per-IO information necessary + * for both the lower (SLI) and upper + * layers (efct). + */ +struct efct_hw_io { + /* Owned by HW */ + + /* reference counter and callback function */ + struct kref ref; + void (*release)(struct kref *arg); + /* used for busy, wait_free, free lists */ + struct list_head list_entry; + /* used for timed_wqe list */ + struct list_head wqe_link; + /* used for io posted dnrx list */ + struct list_head dnrx_link; + /* state of IO: free, busy, wait_free */ + enum efct_hw_io_state state; + /* Work queue object, with link for pending */ + struct efct_hw_wqe wqe; + /* pointer back to hardware context */ + struct efct_hw *hw; + struct efc_remote_node *rnode; + struct efc_dma xfer_rdy; + u16 type; + /* WQ assigned to the exchange */ + struct hw_wq *wq; + /* Exchange is active in FW */ + bool xbusy; + /* Function called on IO completion */ + int + (*done)(struct efct_hw_io *hio, + struct efc_remote_node *rnode, + u32 len, int status, + u32 ext, void *ul_arg); + /* argument passed to "IO done" callback */ + void *arg; + /* Function called on abort completion */ + int + (*abort_done)(struct efct_hw_io *hio, + struct efc_remote_node *rnode, + u32 len, int status, + u32 ext, void *ul_arg); + /* argument passed to "abort done" callback */ + void *abort_arg; + /* needed for bug O127585: length of IO */ + size_t length; + /* timeout value for target WQEs */ + u8 tgt_wqe_timeout; + /* timestamp when current WQE was submitted */ + u64 submit_ticks; + + /* if TRUE, latched status shld be returned */ + bool status_saved; + /* if TRUE, abort is in progress */ + bool abort_in_progress; + u32 saved_status; + u32 saved_len; + u32 saved_ext; + + /* EQ that this HIO came up on */ + struct hw_eq *eq; + /* WQ steering mode request */ + enum efct_hw_wq_steering wq_steering; + /* WQ class if steering mode is Class */ + u8 wq_class; + + /* request tag for this HW IO */ + u16 reqtag; + /* request tag for an abort of this HW IO + * (note: this is a 32 bit value + * to allow us to use UINT32_MAX as an uninitialized value) + */ + u32 abort_reqtag; + u32 indicator; /* XRI */ + struct efc_dma def_sgl; /* default SGL*/ + /* Count of SGEs in default SGL */ + u32 def_sgl_count; + /* pointer to current active SGL */ + struct efc_dma *sgl; + u32 sgl_count; /* count of SGEs in io->sgl */ + u32 first_data_sge; /* index of first data SGE */ + struct efc_dma *ovfl_sgl; /* overflow SGL */ + u32 ovfl_sgl_count; + /* pointer to overflow segment len */ + struct sli4_lsp_sge *ovfl_lsp; + u32 n_sge; /* number of active SGEs */ + u32 sge_offset; + + /* where upper layer can store ref to its IO */ + void *ul_io; +}; + + +/* Typedef for HW "done" callback */ +typedef int (*efct_hw_done_t)(struct efct_hw_io *, struct efc_remote_node *, + u32 len, int status, u32 ext, void *ul_arg); + +enum efct_hw_port { + EFCT_HW_PORT_INIT, + EFCT_HW_PORT_SHUTDOWN, +}; + +/* Node group rpi reference */ +struct efct_hw_rpi_ref { + atomic_t rpi_count; + atomic_t rpi_attached; +}; + +enum efct_hw_link_stat { + EFCT_HW_LINK_STAT_LINK_FAILURE_COUNT, + EFCT_HW_LINK_STAT_LOSS_OF_SYNC_COUNT, + EFCT_HW_LINK_STAT_LOSS_OF_SIGNAL_COUNT, + EFCT_HW_LINK_STAT_PRIMITIVE_SEQ_COUNT, + EFCT_HW_LINK_STAT_INVALID_XMIT_WORD_COUNT, + EFCT_HW_LINK_STAT_CRC_COUNT, + EFCT_HW_LINK_STAT_PRIMITIVE_SEQ_TIMEOUT_COUNT, + EFCT_HW_LINK_STAT_ELASTIC_BUFFER_OVERRUN_COUNT, + EFCT_HW_LINK_STAT_ARB_TIMEOUT_COUNT, + EFCT_HW_LINK_STAT_ADVERTISED_RCV_B2B_CREDIT, + EFCT_HW_LINK_STAT_CURR_RCV_B2B_CREDIT, + EFCT_HW_LINK_STAT_ADVERTISED_XMIT_B2B_CREDIT, + EFCT_HW_LINK_STAT_CURR_XMIT_B2B_CREDIT, + EFCT_HW_LINK_STAT_RCV_EOFA_COUNT, + EFCT_HW_LINK_STAT_RCV_EOFDTI_COUNT, + EFCT_HW_LINK_STAT_RCV_EOFNI_COUNT, + EFCT_HW_LINK_STAT_RCV_SOFF_COUNT, + EFCT_HW_LINK_STAT_RCV_DROPPED_NO_AER_COUNT, + EFCT_HW_LINK_STAT_RCV_DROPPED_NO_RPI_COUNT, + EFCT_HW_LINK_STAT_RCV_DROPPED_NO_XRI_COUNT, + EFCT_HW_LINK_STAT_MAX, +}; + +enum efct_hw_host_stat { + EFCT_HW_HOST_STAT_TX_KBYTE_COUNT, + EFCT_HW_HOST_STAT_RX_KBYTE_COUNT, + EFCT_HW_HOST_STAT_TX_FRAME_COUNT, + EFCT_HW_HOST_STAT_RX_FRAME_COUNT, + EFCT_HW_HOST_STAT_TX_SEQ_COUNT, + EFCT_HW_HOST_STAT_RX_SEQ_COUNT, + EFCT_HW_HOST_STAT_TOTAL_EXCH_ORIG, + EFCT_HW_HOST_STAT_TOTAL_EXCH_RESP, + EFCT_HW_HOSY_STAT_RX_P_BSY_COUNT, + EFCT_HW_HOST_STAT_RX_F_BSY_COUNT, + EFCT_HW_HOST_STAT_DROP_FRM_DUE_TO_NO_RQ_BUF_COUNT, + EFCT_HW_HOST_STAT_EMPTY_RQ_TIMEOUT_COUNT, + EFCT_HW_HOST_STAT_DROP_FRM_DUE_TO_NO_XRI_COUNT, + EFCT_HW_HOST_STAT_EMPTY_XRI_POOL_COUNT, + EFCT_HW_HOST_STAT_MAX, +}; + +enum efct_hw_state { + EFCT_HW_STATE_UNINITIALIZED, + EFCT_HW_STATE_QUEUES_ALLOCATED, + EFCT_HW_STATE_ACTIVE, + EFCT_HW_STATE_RESET_IN_PROGRESS, + EFCT_HW_STATE_TEARDOWN_IN_PROGRESS, +}; + +struct efct_hw_link_stat_counts { + u8 overflow; + u32 counter; +}; + +struct efct_hw_host_stat_counts { + u32 counter; +}; + +#include "efct_hw_queues.h" + +/* Structure used for the hash lookup of queue IDs */ +struct efct_queue_hash { + bool in_use; + u16 id; + u16 index; +}; + +/* WQ callback object */ +struct hw_wq_callback { + u16 instance_index; /* use for request tag */ + void (*callback)(void *arg, u8 *cqe, int status); + void *arg; +}; + +struct efct_hw_config { + u32 n_eq; + u32 n_cq; + u32 n_mq; + u32 n_rq; + u32 n_wq; + u32 n_io; + u32 n_sgl; + u32 speed; + u32 topology; + /* size of the buffers for first burst */ + u32 rq_default_buffer_size; + u8 esoc; + /* The seed for the DIF CRC calculation */ + u16 dif_seed; + u8 dif_mode; + /* Enable driver target wqe timeouts */ + u8 emulate_tgt_wqe_timeout; + bool bounce; + /* Queue topology string */ + const char *queue_topology; + /* MRQ RQ selection policy */ + u8 rq_selection_policy; + /* RQ quanta if rq_selection_policy == 2 */ + u8 rr_quanta; + u32 filter_def[SLI4_CMD_REG_FCFI_NUM_RQ_CFG]; +}; + +struct efct_hw { + struct efct *os; + struct sli4 sli; + u16 ulp_start; + u16 ulp_max; + u32 dump_size; + enum efct_hw_state state; + bool hw_setup_called; + u8 sliport_healthcheck; + u16 watchdog_timeout; + + /* HW configuration, subject to efct_hw_set() */ + struct efct_hw_config config; + + /* calculated queue sizes for each type */ + u32 num_qentries[SLI_QTYPE_MAX]; + + /* Storage for SLI queue objects */ + struct sli4_queue wq[EFCT_HW_MAX_NUM_WQ]; + struct sli4_queue rq[EFCT_HW_MAX_NUM_RQ]; + u16 hw_rq_lookup[EFCT_HW_MAX_NUM_RQ]; + struct sli4_queue mq[EFCT_HW_MAX_NUM_MQ]; + struct sli4_queue cq[EFCT_HW_MAX_NUM_CQ]; + struct sli4_queue eq[EFCT_HW_MAX_NUM_EQ]; + + /* HW queue */ + u32 eq_count; + u32 cq_count; + u32 mq_count; + u32 wq_count; + u32 rq_count; + struct list_head eq_list; + + struct efct_queue_hash cq_hash[EFCT_HW_Q_HASH_SIZE]; + struct efct_queue_hash rq_hash[EFCT_HW_Q_HASH_SIZE]; + struct efct_queue_hash wq_hash[EFCT_HW_Q_HASH_SIZE]; + + /* Storage for HW queue objects */ + struct hw_wq *hw_wq[EFCT_HW_MAX_NUM_WQ]; + struct hw_rq *hw_rq[EFCT_HW_MAX_NUM_RQ]; + struct hw_mq *hw_mq[EFCT_HW_MAX_NUM_MQ]; + struct hw_cq *hw_cq[EFCT_HW_MAX_NUM_CQ]; + struct hw_eq *hw_eq[EFCT_HW_MAX_NUM_EQ]; + /* count of hw_rq[] entries */ + u32 hw_rq_count; + /* count of multirq RQs */ + u32 hw_mrq_count; + + /* pool per class WQs */ + struct efct_varray *wq_class_array[EFCT_HW_MAX_WQ_CLASS]; + /* pool per CPU WQs */ + struct efct_varray *wq_cpu_array[EFCT_HW_MAX_WQ_CPU]; + + /* Sequence objects used in incoming frame processing */ + struct efct_array *seq_pool; + + /* Maintain an ordered, linked list of outstanding HW commands. */ + spinlock_t cmd_lock; + struct list_head cmd_head; + struct list_head cmd_pending; + u32 cmd_head_count; + + struct sli4_link_event link; + struct efc_domain *domain; + + u16 fcf_indicator; + + /* pointer array of IO objects */ + struct efct_hw_io **io; + /* array of WQE buffs mapped to IO objects */ + u8 *wqe_buffs; + + /* IO lock to synchronize list access */ + spinlock_t io_lock; + /* IO lock to synchronize IO aborting */ + spinlock_t io_abort_lock; + /* List of IO objects in use */ + struct list_head io_inuse; + /* List of IO objects with a timed target WQE */ + struct list_head io_timed_wqe; + /* List of IO objects waiting to be freed */ + struct list_head io_wait_free; + /* List of IO objects available for allocation */ + struct list_head io_free; + + struct efc_dma loop_map; + + struct efc_dma xfer_rdy; + + struct efc_dma dump_sges; + + struct efc_dma rnode_mem; + + struct efct_hw_rpi_ref *rpi_ref; + + atomic_t io_alloc_failed_count; + + struct efct_hw_qtop *qtop; + + /* stat: wq sumbit count */ + u32 tcmd_wq_submit[EFCT_HW_MAX_NUM_WQ]; + /* stat: wq complete count */ + u32 tcmd_wq_complete[EFCT_HW_MAX_NUM_WQ]; + /* Timer to periodically check for WQE timeouts */ + struct timer_list wqe_timer; + /* Timer for heartbeat */ + struct timer_list watchdog_timer; + bool in_active_wqe_timer; + bool active_wqe_timer_shutdown; + + struct efct_pool *wq_reqtag_pool; + atomic_t send_frame_seq_id; +}; + +enum efct_hw_io_count_type { + EFCT_HW_IO_INUSE_COUNT, + EFCT_HW_IO_FREE_COUNT, + EFCT_HW_IO_WAIT_FREE_COUNT, + EFCT_HW_IO_N_TOTAL_IO_COUNT, +}; + +/* HW queue data structures */ +struct hw_eq { + struct list_head list_entry; + enum sli4_qtype type; + u32 instance; + u32 entry_count; + u32 entry_size; + struct efct_hw *hw; + struct sli4_queue *queue; + struct list_head cq_list; + u32 use_count; + struct efct_varray *wq_array; +}; + +struct hw_cq { + struct list_head list_entry; + enum sli4_qtype type; + u32 instance; + u32 entry_count; + u32 entry_size; + struct hw_eq *eq; + struct sli4_queue *queue; + struct list_head q_list; + u32 use_count; +}; + +struct hw_q { + struct list_head list_entry; + enum sli4_qtype type; +}; + +struct hw_mq { + struct list_head list_entry; + enum sli4_qtype type; + u32 instance; + + u32 entry_count; + u32 entry_size; + struct hw_cq *cq; + struct sli4_queue *queue; + + u32 use_count; +}; + +struct hw_wq { + struct list_head list_entry; + enum sli4_qtype type; + u32 instance; + struct efct_hw *hw; + + u32 entry_count; + u32 entry_size; + struct hw_cq *cq; + struct sli4_queue *queue; + u32 class; + u8 ulp; + + /* WQ consumed */ + u32 wqec_set_count; + u32 wqec_count; + u32 free_count; + u32 total_submit_count; + struct list_head pending_list; + + /* HW IO allocated for use with Send Frame */ + struct efct_hw_io *send_frame_io; + + /* Stats */ + u32 use_count; + u32 wq_pending_count; +}; + +struct hw_rq { + struct list_head list_entry; + enum sli4_qtype type; + u32 instance; + + u32 entry_count; + u32 use_count; + u32 hdr_entry_size; + u32 first_burst_entry_size; + u32 data_entry_size; + u8 ulp; + bool is_mrq; + u32 base_mrq_id; + + struct hw_cq *cq; + + u8 filter_mask; + struct sli4_queue *hdr; + struct sli4_queue *first_burst; + struct sli4_queue *data; + + struct efc_hw_rq_buffer *hdr_buf; + struct efc_hw_rq_buffer *fb_buf; + struct efc_hw_rq_buffer *payload_buf; + /* RQ tracker for this RQ */ + struct efc_hw_sequence **rq_tracker; +}; + +struct efct_hw_global { + const char *queue_topology_string; +}; + +extern struct efct_hw_global hw_global; + +struct efct_hw_send_frame_context { + struct efct_hw *hw; + struct hw_wq_callback *wqcb; + struct efct_hw_wqe wqe; + void (*callback)(int status, void *arg); + void *arg; + + /* General purpose elements */ + struct efc_hw_sequence *seq; + struct efc_dma payload; +}; + +#define EFCT_HW_OBJECT_G5 0xfeaa0001 +#define EFCT_HW_OBJECT_G6 0xfeaa0003 +struct efct_hw_grp_hdr { + u32 size; + __be32 magic_number; + u32 word2; + u8 rev_name[128]; + u8 date[12]; + u8 revision[32]; +}; + +#endif /* __EFCT_H__ */ -- 2.13.7