Re: [PATCH for-next 1/2] RDMA/hns: Refactor eq code for hip06

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The cover-letter is there. There is no need to resend.
Sorry again.

Eason

On 2017/11/14 16:53, Liuyixian (Eason) wrote:
> Sorry, cover-letter has been lost for some unknown problem.
> I will resend the patch set.
> 
> On 2017/11/14 17:26, Yixian Liu wrote:
>> Considering the compatibility of supporting hip08's eq
>> process and possible changes of data structure, this patch
>> refactors the eq code structure of hip06.
>>
>> We move all the eq process code for hip06 from hns_roce_eq.c
>> into hns_roce_hw_v1.c, and also for hns_roce_eq.h. With
>> these changes, it will be convenient to add the eq support
>> for later hardware version.
>>
>> Signed-off-by: Yixian Liu <liuyixian@xxxxxxxxxx>
>> Reviewed-by: Lijun Ou <oulijun@xxxxxxxxxx>
>> Reviewed-by: Wei Hu (Xavier) <xavier.huwei@xxxxxxxxxx>
>> ---
>>  drivers/infiniband/hw/hns/Makefile          |   2 +-
>>  drivers/infiniband/hw/hns/hns_roce_cmd.c    |   1 +
>>  drivers/infiniband/hw/hns/hns_roce_cq.c     |  19 +-
>>  drivers/infiniband/hw/hns/hns_roce_device.h |  57 ++-
>>  drivers/infiniband/hw/hns/hns_roce_eq.c     | 759 ----------------------------
>>  drivers/infiniband/hw/hns/hns_roce_eq.h     | 134 -----
>>  drivers/infiniband/hw/hns/hns_roce_hw_v1.c  | 740 ++++++++++++++++++++++++++-
>>  drivers/infiniband/hw/hns/hns_roce_hw_v1.h  |  44 +-
>>  drivers/infiniband/hw/hns/hns_roce_main.c   |  16 +-
>>  drivers/infiniband/hw/hns/hns_roce_qp.c     |   1 +
>>  10 files changed, 843 insertions(+), 930 deletions(-)
>>  delete mode 100644 drivers/infiniband/hw/hns/hns_roce_eq.c
>>  delete mode 100644 drivers/infiniband/hw/hns/hns_roce_eq.h
>>
>> diff --git a/drivers/infiniband/hw/hns/Makefile b/drivers/infiniband/hw/hns/Makefile
>> index ff426a6..97bf2cd 100644
>> --- a/drivers/infiniband/hw/hns/Makefile
>> +++ b/drivers/infiniband/hw/hns/Makefile
>> @@ -5,7 +5,7 @@
>>  ccflags-y :=  -Idrivers/net/ethernet/hisilicon/hns3
>>  
>>  obj-$(CONFIG_INFINIBAND_HNS) += hns-roce.o
>> -hns-roce-objs := hns_roce_main.o hns_roce_cmd.o hns_roce_eq.o hns_roce_pd.o \
>> +hns-roce-objs := hns_roce_main.o hns_roce_cmd.o hns_roce_pd.o \
>>  	hns_roce_ah.o hns_roce_hem.o hns_roce_mr.o hns_roce_qp.o \
>>  	hns_roce_cq.o hns_roce_alloc.o
>>  obj-$(CONFIG_INFINIBAND_HNS_HIP06) += hns-roce-hw-v1.o
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_cmd.c b/drivers/infiniband/hw/hns/hns_roce_cmd.c
>> index 1085cb2..9ebe839 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_cmd.c
>> +++ b/drivers/infiniband/hw/hns/hns_roce_cmd.c
>> @@ -103,6 +103,7 @@ void hns_roce_cmd_event(struct hns_roce_dev *hr_dev, u16 token, u8 status,
>>  	context->out_param = out_param;
>>  	complete(&context->done);
>>  }
>> +EXPORT_SYMBOL_GPL(hns_roce_cmd_event);
>>  
>>  /* this should be called with "use_events" */
>>  static int __hns_roce_cmd_mbox_wait(struct hns_roce_dev *hr_dev, u64 in_param,
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_cq.c b/drivers/infiniband/hw/hns/hns_roce_cq.c
>> index 2111b57..bccc9b5 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_cq.c
>> +++ b/drivers/infiniband/hw/hns/hns_roce_cq.c
>> @@ -196,15 +196,14 @@ void hns_roce_free_cq(struct hns_roce_dev *hr_dev, struct hns_roce_cq *hr_cq)
>>  	if (ret)
>>  		dev_err(dev, "HW2SW_CQ failed (%d) for CQN %06lx\n", ret,
>>  			hr_cq->cqn);
>> -	if (hr_dev->eq_table.eq) {
>> -		/* Waiting interrupt process procedure carried out */
>> -		synchronize_irq(hr_dev->eq_table.eq[hr_cq->vector].irq);
>> -
>> -		/* wait for all interrupt processed */
>> -		if (atomic_dec_and_test(&hr_cq->refcount))
>> -			complete(&hr_cq->free);
>> -		wait_for_completion(&hr_cq->free);
>> -	}
>> +
>> +	/* Waiting interrupt process procedure carried out */
>> +	synchronize_irq(hr_dev->eq_table.eq[hr_cq->vector].irq);
>> +
>> +	/* wait for all interrupt processed */
>> +	if (atomic_dec_and_test(&hr_cq->refcount))
>> +		complete(&hr_cq->free);
>> +	wait_for_completion(&hr_cq->free);
>>  
>>  	spin_lock_irq(&cq_table->lock);
>>  	radix_tree_delete(&cq_table->tree, hr_cq->cqn);
>> @@ -460,6 +459,7 @@ void hns_roce_cq_completion(struct hns_roce_dev *hr_dev, u32 cqn)
>>  	++cq->arm_sn;
>>  	cq->comp(cq);
>>  }
>> +EXPORT_SYMBOL_GPL(hns_roce_cq_completion);
>>  
>>  void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type)
>>  {
>> @@ -482,6 +482,7 @@ void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type)
>>  	if (atomic_dec_and_test(&cq->refcount))
>>  		complete(&cq->free);
>>  }
>> +EXPORT_SYMBOL_GPL(hns_roce_cq_event);
>>  
>>  int hns_roce_init_cq_table(struct hns_roce_dev *hr_dev)
>>  {
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_device.h b/drivers/infiniband/hw/hns/hns_roce_device.h
>> index 01d3d69..9aa9e94 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_device.h
>> +++ b/drivers/infiniband/hw/hns/hns_roce_device.h
>> @@ -62,12 +62,16 @@
>>  #define HNS_ROCE_CQE_WCMD_EMPTY_BIT		0x2
>>  #define HNS_ROCE_MIN_CQE_CNT			16
>>  
>> -#define HNS_ROCE_MAX_IRQ_NUM			34
>> +#define HNS_ROCE_MAX_IRQ_NUM			128
>>  
>> -#define HNS_ROCE_COMP_VEC_NUM			32
>> +#define EQ_ENABLE				1
>> +#define EQ_DISABLE				0
>>  
>> -#define HNS_ROCE_AEQE_VEC_NUM			1
>> -#define HNS_ROCE_AEQE_OF_VEC_NUM		1
>> +#define HNS_ROCE_CEQ				0
>> +#define HNS_ROCE_AEQ				1
>> +
>> +#define HNS_ROCE_CEQ_ENTRY_SIZE			0x4
>> +#define HNS_ROCE_AEQ_ENTRY_SIZE			0x10
>>  
>>  /* 4G/4K = 1M */
>>  #define HNS_ROCE_SL_SHIFT			28
>> @@ -485,6 +489,45 @@ struct hns_roce_ib_iboe {
>>  	u8			phy_port[HNS_ROCE_MAX_PORTS];
>>  };
>>  
>> +enum {
>> +	HNS_ROCE_EQ_STAT_INVALID  = 0,
>> +	HNS_ROCE_EQ_STAT_VALID    = 2,
>> +};
>> +
>> +struct hns_roce_ceqe {
>> +	u32			comp;
>> +};
>> +
>> +struct hns_roce_aeqe {
>> +	u32 asyn;
>> +	union {
>> +		struct {
>> +			u32 qp;
>> +			u32 rsv0;
>> +			u32 rsv1;
>> +		} qp_event;
>> +
>> +		struct {
>> +			u32 cq;
>> +			u32 rsv0;
>> +			u32 rsv1;
>> +		} cq_event;
>> +
>> +		struct {
>> +			u32 ceqe;
>> +			u32 rsv0;
>> +			u32 rsv1;
>> +		} ce_event;
>> +
>> +		struct {
>> +			__le64  out_param;
>> +			__le16  token;
>> +			u8	status;
>> +			u8	rsv0;
>> +		} __packed cmd;
>> +	 } event;
>> +};
>> +
>>  struct hns_roce_eq {
>>  	struct hns_roce_dev		*hr_dev;
>>  	void __iomem			*doorbell;
>> @@ -502,7 +545,7 @@ struct hns_roce_eq {
>>  
>>  struct hns_roce_eq_table {
>>  	struct hns_roce_eq	*eq;
>> -	void __iomem		**eqc_base;
>> +	void __iomem		**eqc_base; /* only for hw v1 */
>>  };
>>  
>>  struct hns_roce_caps {
>> @@ -550,7 +593,7 @@ struct hns_roce_caps {
>>  	u32		pbl_buf_pg_sz;
>>  	u32		pbl_hop_num;
>>  	int		aeqe_depth;
>> -	int		ceqe_depth[HNS_ROCE_COMP_VEC_NUM];
>> +	int		ceqe_depth;
>>  	enum ib_mtu	max_mtu;
>>  	u32		qpc_bt_num;
>>  	u32		srqc_bt_num;
>> @@ -623,6 +666,8 @@ struct hns_roce_hw {
>>  	int (*dereg_mr)(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr);
>>  	int (*destroy_cq)(struct ib_cq *ibcq);
>>  	int (*modify_cq)(struct ib_cq *cq, u16 cq_count, u16 cq_period);
>> +	int (*init_eq)(struct hns_roce_dev *hr_dev);
>> +	void (*cleanup_eq)(struct hns_roce_dev *hr_dev);
>>  };
>>  
>>  struct hns_roce_dev {
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_eq.c b/drivers/infiniband/hw/hns/hns_roce_eq.c
>> deleted file mode 100644
>> index d184431..0000000
>> --- a/drivers/infiniband/hw/hns/hns_roce_eq.c
>> +++ /dev/null
>> @@ -1,759 +0,0 @@
>> -/*
>> - * Copyright (c) 2016 Hisilicon Limited.
>> - *
>> - * This software is available to you under a choice of one of two
>> - * licenses.  You may choose to be licensed under the terms of the GNU
>> - * General Public License (GPL) Version 2, available from the file
>> - * COPYING in the main directory of this source tree, or the
>> - * OpenIB.org BSD license below:
>> - *
>> - *     Redistribution and use in source and binary forms, with or
>> - *     without modification, are permitted provided that the following
>> - *     conditions are met:
>> - *
>> - *      - Redistributions of source code must retain the above
>> - *        copyright notice, this list of conditions and the following
>> - *        disclaimer.
>> - *
>> - *      - Redistributions in binary form must reproduce the above
>> - *        copyright notice, this list of conditions and the following
>> - *        disclaimer in the documentation and/or other materials
>> - *        provided with the distribution.
>> - *
>> - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
>> - * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
>> - * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
>> - * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
>> - * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
>> - * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
>> - * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
>> - * SOFTWARE.
>> - */
>> -
>> -#include <linux/platform_device.h>
>> -#include <linux/interrupt.h>
>> -#include "hns_roce_common.h"
>> -#include "hns_roce_device.h"
>> -#include "hns_roce_eq.h"
>> -
>> -static void eq_set_cons_index(struct hns_roce_eq *eq, int req_not)
>> -{
>> -	roce_raw_write((eq->cons_index & CONS_INDEX_MASK) |
>> -		      (req_not << eq->log_entries), eq->doorbell);
>> -	/* Memory barrier */
>> -	mb();
>> -}
>> -
>> -static struct hns_roce_aeqe *get_aeqe(struct hns_roce_eq *eq, u32 entry)
>> -{
>> -	unsigned long off = (entry & (eq->entries - 1)) *
>> -			     HNS_ROCE_AEQ_ENTRY_SIZE;
>> -
>> -	return (struct hns_roce_aeqe *)((u8 *)
>> -		(eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
>> -		off % HNS_ROCE_BA_SIZE);
>> -}
>> -
>> -static struct hns_roce_aeqe *next_aeqe_sw(struct hns_roce_eq *eq)
>> -{
>> -	struct hns_roce_aeqe *aeqe = get_aeqe(eq, eq->cons_index);
>> -
>> -	return (roce_get_bit(aeqe->asyn, HNS_ROCE_AEQE_U32_4_OWNER_S) ^
>> -		!!(eq->cons_index & eq->entries)) ? aeqe : NULL;
>> -}
>> -
>> -static void hns_roce_wq_catas_err_handle(struct hns_roce_dev *hr_dev,
>> -					 struct hns_roce_aeqe *aeqe, int qpn)
>> -{
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -
>> -	dev_warn(dev, "Local Work Queue Catastrophic Error.\n");
>> -	switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> -			       HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> -	case HNS_ROCE_LWQCE_QPC_ERROR:
>> -		dev_warn(dev, "QP %d, QPC error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LWQCE_MTU_ERROR:
>> -		dev_warn(dev, "QP %d, MTU error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LWQCE_WQE_BA_ADDR_ERROR:
>> -		dev_warn(dev, "QP %d, WQE BA addr error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LWQCE_WQE_ADDR_ERROR:
>> -		dev_warn(dev, "QP %d, WQE addr error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LWQCE_SQ_WQE_SHIFT_ERROR:
>> -		dev_warn(dev, "QP %d, WQE shift error\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LWQCE_SL_ERROR:
>> -		dev_warn(dev, "QP %d, SL error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LWQCE_PORT_ERROR:
>> -		dev_warn(dev, "QP %d, port error.\n", qpn);
>> -		break;
>> -	default:
>> -		break;
>> -	}
>> -}
>> -
>> -static void hns_roce_local_wq_access_err_handle(struct hns_roce_dev *hr_dev,
>> -						struct hns_roce_aeqe *aeqe,
>> -						int qpn)
>> -{
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -
>> -	dev_warn(dev, "Local Access Violation Work Queue Error.\n");
>> -	switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> -			       HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> -	case HNS_ROCE_LAVWQE_R_KEY_VIOLATION:
>> -		dev_warn(dev, "QP %d, R_key violation.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LAVWQE_LENGTH_ERROR:
>> -		dev_warn(dev, "QP %d, length error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LAVWQE_VA_ERROR:
>> -		dev_warn(dev, "QP %d, VA error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LAVWQE_PD_ERROR:
>> -		dev_err(dev, "QP %d, PD error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LAVWQE_RW_ACC_ERROR:
>> -		dev_warn(dev, "QP %d, rw acc error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LAVWQE_KEY_STATE_ERROR:
>> -		dev_warn(dev, "QP %d, key state error.\n", qpn);
>> -		break;
>> -	case HNS_ROCE_LAVWQE_MR_OPERATION_ERROR:
>> -		dev_warn(dev, "QP %d, MR operation error.\n", qpn);
>> -		break;
>> -	default:
>> -		break;
>> -	}
>> -}
>> -
>> -static void hns_roce_qp_err_handle(struct hns_roce_dev *hr_dev,
>> -				   struct hns_roce_aeqe *aeqe,
>> -				   int event_type)
>> -{
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -	int phy_port;
>> -	int qpn;
>> -
>> -	qpn = roce_get_field(aeqe->event.qp_event.qp,
>> -			     HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> -			     HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S);
>> -	phy_port = roce_get_field(aeqe->event.qp_event.qp,
>> -			HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M,
>> -			HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S);
>> -	if (qpn <= 1)
>> -		qpn = HNS_ROCE_MAX_PORTS * qpn + phy_port;
>> -
>> -	switch (event_type) {
>> -	case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
>> -		dev_warn(dev, "Invalid Req Local Work Queue Error.\n"
>> -			      "QP %d, phy_port %d.\n", qpn, phy_port);
>> -		break;
>> -	case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
>> -		hns_roce_wq_catas_err_handle(hr_dev, aeqe, qpn);
>> -		break;
>> -	case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
>> -		hns_roce_local_wq_access_err_handle(hr_dev, aeqe, qpn);
>> -		break;
>> -	default:
>> -		break;
>> -	}
>> -
>> -	hns_roce_qp_event(hr_dev, qpn, event_type);
>> -}
>> -
>> -static void hns_roce_cq_err_handle(struct hns_roce_dev *hr_dev,
>> -				   struct hns_roce_aeqe *aeqe,
>> -				   int event_type)
>> -{
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -	u32 cqn;
>> -
>> -	cqn = le32_to_cpu(roce_get_field(aeqe->event.cq_event.cq,
>> -		    HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
>> -		    HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S));
>> -
>> -	switch (event_type) {
>> -	case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
>> -		dev_warn(dev, "CQ 0x%x access err.\n", cqn);
>> -		break;
>> -	case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
>> -		dev_warn(dev, "CQ 0x%x overflow\n", cqn);
>> -		break;
>> -	case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
>> -		dev_warn(dev, "CQ 0x%x ID invalid.\n", cqn);
>> -		break;
>> -	default:
>> -		break;
>> -	}
>> -
>> -	hns_roce_cq_event(hr_dev, cqn, event_type);
>> -}
>> -
>> -static void hns_roce_db_overflow_handle(struct hns_roce_dev *hr_dev,
>> -					struct hns_roce_aeqe *aeqe)
>> -{
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -
>> -	switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> -			       HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> -	case HNS_ROCE_DB_SUBTYPE_SDB_OVF:
>> -		dev_warn(dev, "SDB overflow.\n");
>> -		break;
>> -	case HNS_ROCE_DB_SUBTYPE_SDB_ALM_OVF:
>> -		dev_warn(dev, "SDB almost overflow.\n");
>> -		break;
>> -	case HNS_ROCE_DB_SUBTYPE_SDB_ALM_EMP:
>> -		dev_warn(dev, "SDB almost empty.\n");
>> -		break;
>> -	case HNS_ROCE_DB_SUBTYPE_ODB_OVF:
>> -		dev_warn(dev, "ODB overflow.\n");
>> -		break;
>> -	case HNS_ROCE_DB_SUBTYPE_ODB_ALM_OVF:
>> -		dev_warn(dev, "ODB almost overflow.\n");
>> -		break;
>> -	case HNS_ROCE_DB_SUBTYPE_ODB_ALM_EMP:
>> -		dev_warn(dev, "SDB almost empty.\n");
>> -		break;
>> -	default:
>> -		break;
>> -	}
>> -}
>> -
>> -static int hns_roce_aeq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
>> -{
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -	struct hns_roce_aeqe *aeqe;
>> -	int aeqes_found = 0;
>> -	int event_type;
>> -
>> -	while ((aeqe = next_aeqe_sw(eq))) {
>> -		dev_dbg(dev, "aeqe = %p, aeqe->asyn.event_type = 0x%lx\n", aeqe,
>> -			roce_get_field(aeqe->asyn,
>> -				       HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> -				       HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> -		/* Memory barrier */
>> -		rmb();
>> -
>> -		event_type = roce_get_field(aeqe->asyn,
>> -				HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> -				HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S);
>> -		switch (event_type) {
>> -		case HNS_ROCE_EVENT_TYPE_PATH_MIG:
>> -			dev_warn(dev, "PATH MIG not supported\n");
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_COMM_EST:
>> -			dev_warn(dev, "COMMUNICATION established\n");
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_SQ_DRAINED:
>> -			dev_warn(dev, "SQ DRAINED not supported\n");
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_PATH_MIG_FAILED:
>> -			dev_warn(dev, "PATH MIG failed\n");
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
>> -		case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
>> -		case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
>> -			hns_roce_qp_err_handle(hr_dev, aeqe, event_type);
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_SRQ_LIMIT_REACH:
>> -		case HNS_ROCE_EVENT_TYPE_SRQ_CATAS_ERROR:
>> -		case HNS_ROCE_EVENT_TYPE_SRQ_LAST_WQE_REACH:
>> -			dev_warn(dev, "SRQ not support!\n");
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
>> -		case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
>> -		case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
>> -			hns_roce_cq_err_handle(hr_dev, aeqe, event_type);
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_PORT_CHANGE:
>> -			dev_warn(dev, "port change.\n");
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_MB:
>> -			hns_roce_cmd_event(hr_dev,
>> -					   le16_to_cpu(aeqe->event.cmd.token),
>> -					   aeqe->event.cmd.status,
>> -					   le64_to_cpu(aeqe->event.cmd.out_param
>> -					   ));
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_DB_OVERFLOW:
>> -			hns_roce_db_overflow_handle(hr_dev, aeqe);
>> -			break;
>> -		case HNS_ROCE_EVENT_TYPE_CEQ_OVERFLOW:
>> -			dev_warn(dev, "CEQ 0x%lx overflow.\n",
>> -			roce_get_field(aeqe->event.ce_event.ceqe,
>> -				     HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M,
>> -				     HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S));
>> -			break;
>> -		default:
>> -			dev_warn(dev, "Unhandled event %d on EQ %d at index %u\n",
>> -				 event_type, eq->eqn, eq->cons_index);
>> -			break;
>> -		}
>> -
>> -		eq->cons_index++;
>> -		aeqes_found = 1;
>> -
>> -		if (eq->cons_index > 2 * hr_dev->caps.aeqe_depth - 1) {
>> -			dev_warn(dev, "cons_index overflow, set back to zero\n"
>> -				);
>> -			eq->cons_index = 0;
>> -		}
>> -	}
>> -
>> -	eq_set_cons_index(eq, 0);
>> -
>> -	return aeqes_found;
>> -}
>> -
>> -static struct hns_roce_ceqe *get_ceqe(struct hns_roce_eq *eq, u32 entry)
>> -{
>> -	unsigned long off = (entry & (eq->entries - 1)) *
>> -			     HNS_ROCE_CEQ_ENTRY_SIZE;
>> -
>> -	return (struct hns_roce_ceqe *)((u8 *)
>> -			(eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
>> -			off % HNS_ROCE_BA_SIZE);
>> -}
>> -
>> -static struct hns_roce_ceqe *next_ceqe_sw(struct hns_roce_eq *eq)
>> -{
>> -	struct hns_roce_ceqe *ceqe = get_ceqe(eq, eq->cons_index);
>> -
>> -	return (!!(roce_get_bit(ceqe->ceqe.comp,
>> -		 HNS_ROCE_CEQE_CEQE_COMP_OWNER_S))) ^
>> -		 (!!(eq->cons_index & eq->entries)) ? ceqe : NULL;
>> -}
>> -
>> -static int hns_roce_ceq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
>> -{
>> -	struct hns_roce_ceqe *ceqe;
>> -	int ceqes_found = 0;
>> -	u32 cqn;
>> -
>> -	while ((ceqe = next_ceqe_sw(eq))) {
>> -		/* Memory barrier */
>> -		rmb();
>> -		cqn = roce_get_field(ceqe->ceqe.comp,
>> -				     HNS_ROCE_CEQE_CEQE_COMP_CQN_M,
>> -				     HNS_ROCE_CEQE_CEQE_COMP_CQN_S);
>> -		hns_roce_cq_completion(hr_dev, cqn);
>> -
>> -		++eq->cons_index;
>> -		ceqes_found = 1;
>> -
>> -		if (eq->cons_index > 2 * hr_dev->caps.ceqe_depth[eq->eqn] - 1) {
>> -			dev_warn(&eq->hr_dev->pdev->dev,
>> -				"cons_index overflow, set back to zero\n");
>> -			eq->cons_index = 0;
>> -		}
>> -	}
>> -
>> -	eq_set_cons_index(eq, 0);
>> -
>> -	return ceqes_found;
>> -}
>> -
>> -static int hns_roce_aeq_ovf_int(struct hns_roce_dev *hr_dev,
>> -				struct hns_roce_eq *eq)
>> -{
>> -	struct device *dev = &eq->hr_dev->pdev->dev;
>> -	int eqovf_found = 0;
>> -	u32 caepaemask_val;
>> -	u32 cealmovf_val;
>> -	u32 caepaest_val;
>> -	u32 aeshift_val;
>> -	u32 ceshift_val;
>> -	u32 cemask_val;
>> -	int i = 0;
>> -
>> -	/**
>> -	 * AEQ overflow ECC mult bit err CEQ overflow alarm
>> -	 * must clear interrupt, mask irq, clear irq, cancel mask operation
>> -	 */
>> -	aeshift_val = roce_read(hr_dev, ROCEE_CAEP_AEQC_AEQE_SHIFT_REG);
>> -
>> -	if (roce_get_bit(aeshift_val,
>> -		ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQ_ALM_OVF_INT_ST_S) == 1) {
>> -		dev_warn(dev, "AEQ overflow!\n");
>> -
>> -		/* Set mask */
>> -		caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> -		roce_set_bit(caepaemask_val,
>> -			     ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> -			     HNS_ROCE_INT_MASK_ENABLE);
>> -		roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
>> -
>> -		/* Clear int state(INT_WC : write 1 clear) */
>> -		caepaest_val = roce_read(hr_dev, ROCEE_CAEP_AE_ST_REG);
>> -		roce_set_bit(caepaest_val,
>> -			     ROCEE_CAEP_AE_ST_CAEP_AEQ_ALM_OVF_S, 1);
>> -		roce_write(hr_dev, ROCEE_CAEP_AE_ST_REG, caepaest_val);
>> -
>> -		/* Clear mask */
>> -		caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> -		roce_set_bit(caepaemask_val,
>> -			     ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> -			     HNS_ROCE_INT_MASK_DISABLE);
>> -		roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
>> -	}
>> -
>> -	/* CEQ almost overflow */
>> -	for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
>> -		ceshift_val = roce_read(hr_dev, ROCEE_CAEP_CEQC_SHIFT_0_REG +
>> -					i * CEQ_REG_OFFSET);
>> -
>> -		if (roce_get_bit(ceshift_val,
>> -		ROCEE_CAEP_CEQC_SHIFT_CAEP_CEQ_ALM_OVF_INT_ST_S) == 1) {
>> -			dev_warn(dev, "CEQ[%d] almost overflow!\n", i);
>> -			eqovf_found++;
>> -
>> -			/* Set mask */
>> -			cemask_val = roce_read(hr_dev,
>> -					       ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> -					       i * CEQ_REG_OFFSET);
>> -			roce_set_bit(cemask_val,
>> -				ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
>> -				HNS_ROCE_INT_MASK_ENABLE);
>> -			roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> -				   i * CEQ_REG_OFFSET, cemask_val);
>> -
>> -			/* Clear int state(INT_WC : write 1 clear) */
>> -			cealmovf_val = roce_read(hr_dev,
>> -				       ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
>> -				       i * CEQ_REG_OFFSET);
>> -			roce_set_bit(cealmovf_val,
>> -				     ROCEE_CAEP_CEQ_ALM_OVF_CAEP_CEQ_ALM_OVF_S,
>> -				     1);
>> -			roce_write(hr_dev, ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
>> -				    i * CEQ_REG_OFFSET, cealmovf_val);
>> -
>> -			/* Clear mask */
>> -			cemask_val = roce_read(hr_dev,
>> -				     ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> -				     i * CEQ_REG_OFFSET);
>> -			roce_set_bit(cemask_val,
>> -			       ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
>> -			       HNS_ROCE_INT_MASK_DISABLE);
>> -			roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> -				   i * CEQ_REG_OFFSET, cemask_val);
>> -		}
>> -	}
>> -
>> -	/* ECC multi-bit error alarm */
>> -	dev_warn(dev, "ECC UCERR ALARM: 0x%x, 0x%x, 0x%x\n",
>> -		 roce_read(hr_dev, ROCEE_ECC_UCERR_ALM0_REG),
>> -		 roce_read(hr_dev, ROCEE_ECC_UCERR_ALM1_REG),
>> -		 roce_read(hr_dev, ROCEE_ECC_UCERR_ALM2_REG));
>> -
>> -	dev_warn(dev, "ECC CERR ALARM: 0x%x, 0x%x, 0x%x\n",
>> -		 roce_read(hr_dev, ROCEE_ECC_CERR_ALM0_REG),
>> -		 roce_read(hr_dev, ROCEE_ECC_CERR_ALM1_REG),
>> -		 roce_read(hr_dev, ROCEE_ECC_CERR_ALM2_REG));
>> -
>> -	return eqovf_found;
>> -}
>> -
>> -static int hns_roce_eq_int(struct hns_roce_dev *hr_dev, struct hns_roce_eq *eq)
>> -{
>> -	int eqes_found = 0;
>> -
>> -	if (likely(eq->type_flag == HNS_ROCE_CEQ))
>> -		/* CEQ irq routine, CEQ is pulse irq, not clear */
>> -		eqes_found = hns_roce_ceq_int(hr_dev, eq);
>> -	else if (likely(eq->type_flag == HNS_ROCE_AEQ))
>> -		/* AEQ irq routine, AEQ is pulse irq, not clear */
>> -		eqes_found = hns_roce_aeq_int(hr_dev, eq);
>> -	else
>> -		/* AEQ queue overflow irq */
>> -		eqes_found = hns_roce_aeq_ovf_int(hr_dev, eq);
>> -
>> -	return eqes_found;
>> -}
>> -
>> -static irqreturn_t hns_roce_msi_x_interrupt(int irq, void *eq_ptr)
>> -{
>> -	int int_work = 0;
>> -	struct hns_roce_eq  *eq  = eq_ptr;
>> -	struct hns_roce_dev *hr_dev = eq->hr_dev;
>> -
>> -	int_work = hns_roce_eq_int(hr_dev, eq);
>> -
>> -	return IRQ_RETVAL(int_work);
>> -}
>> -
>> -static void hns_roce_enable_eq(struct hns_roce_dev *hr_dev, int eq_num,
>> -			       int enable_flag)
>> -{
>> -	void __iomem *eqc = hr_dev->eq_table.eqc_base[eq_num];
>> -	u32 val;
>> -
>> -	val = readl(eqc);
>> -
>> -	if (enable_flag)
>> -		roce_set_field(val,
>> -			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> -			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> -			       HNS_ROCE_EQ_STAT_VALID);
>> -	else
>> -		roce_set_field(val,
>> -			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> -			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> -			       HNS_ROCE_EQ_STAT_INVALID);
>> -	writel(val, eqc);
>> -}
>> -
>> -static int hns_roce_create_eq(struct hns_roce_dev *hr_dev,
>> -			      struct hns_roce_eq *eq)
>> -{
>> -	void __iomem *eqc = hr_dev->eq_table.eqc_base[eq->eqn];
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -	dma_addr_t tmp_dma_addr;
>> -	u32 eqconsindx_val = 0;
>> -	u32 eqcuridx_val = 0;
>> -	u32 eqshift_val = 0;
>> -	int num_bas = 0;
>> -	int ret;
>> -	int i;
>> -
>> -	num_bas = (PAGE_ALIGN(eq->entries * eq->eqe_size) +
>> -		   HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
>> -
>> -	if ((eq->entries * eq->eqe_size) > HNS_ROCE_BA_SIZE) {
>> -		dev_err(dev, "[error]eq buf %d gt ba size(%d) need bas=%d\n",
>> -			(eq->entries * eq->eqe_size), HNS_ROCE_BA_SIZE,
>> -			num_bas);
>> -		return -EINVAL;
>> -	}
>> -
>> -	eq->buf_list = kcalloc(num_bas, sizeof(*eq->buf_list), GFP_KERNEL);
>> -	if (!eq->buf_list)
>> -		return -ENOMEM;
>> -
>> -	for (i = 0; i < num_bas; ++i) {
>> -		eq->buf_list[i].buf = dma_alloc_coherent(dev, HNS_ROCE_BA_SIZE,
>> -							 &tmp_dma_addr,
>> -							 GFP_KERNEL);
>> -		if (!eq->buf_list[i].buf) {
>> -			ret = -ENOMEM;
>> -			goto err_out_free_pages;
>> -		}
>> -
>> -		eq->buf_list[i].map = tmp_dma_addr;
>> -		memset(eq->buf_list[i].buf, 0, HNS_ROCE_BA_SIZE);
>> -	}
>> -	eq->cons_index = 0;
>> -	roce_set_field(eqshift_val,
>> -		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> -		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> -		       HNS_ROCE_EQ_STAT_INVALID);
>> -	roce_set_field(eqshift_val,
>> -		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_M,
>> -		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_S,
>> -		       eq->log_entries);
>> -	writel(eqshift_val, eqc);
>> -
>> -	/* Configure eq extended address 12~44bit */
>> -	writel((u32)(eq->buf_list[0].map >> 12), eqc + 4);
>> -
>> -	/*
>> -	 * Configure eq extended address 45~49 bit.
>> -	 * 44 = 32 + 12, When evaluating addr to hardware, shift 12 because of
>> -	 * using 4K page, and shift more 32 because of
>> -	 * caculating the high 32 bit value evaluated to hardware.
>> -	 */
>> -	roce_set_field(eqcuridx_val, ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_M,
>> -		       ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_S,
>> -		       eq->buf_list[0].map >> 44);
>> -	roce_set_field(eqcuridx_val,
>> -		       ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_M,
>> -		       ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_S, 0);
>> -	writel(eqcuridx_val, eqc + 8);
>> -
>> -	/* Configure eq consumer index */
>> -	roce_set_field(eqconsindx_val,
>> -		       ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_M,
>> -		       ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_S, 0);
>> -	writel(eqconsindx_val, eqc + 0xc);
>> -
>> -	return 0;
>> -
>> -err_out_free_pages:
>> -	for (i = i - 1; i >= 0; i--)
>> -		dma_free_coherent(dev, HNS_ROCE_BA_SIZE, eq->buf_list[i].buf,
>> -				  eq->buf_list[i].map);
>> -
>> -	kfree(eq->buf_list);
>> -	return ret;
>> -}
>> -
>> -static void hns_roce_free_eq(struct hns_roce_dev *hr_dev,
>> -			     struct hns_roce_eq *eq)
>> -{
>> -	int i = 0;
>> -	int npages = (PAGE_ALIGN(eq->eqe_size * eq->entries) +
>> -		      HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
>> -
>> -	if (!eq->buf_list)
>> -		return;
>> -
>> -	for (i = 0; i < npages; ++i)
>> -		dma_free_coherent(&hr_dev->pdev->dev, HNS_ROCE_BA_SIZE,
>> -				  eq->buf_list[i].buf, eq->buf_list[i].map);
>> -
>> -	kfree(eq->buf_list);
>> -}
>> -
>> -static void hns_roce_int_mask_en(struct hns_roce_dev *hr_dev)
>> -{
>> -	int i = 0;
>> -	u32 aemask_val;
>> -	int masken = 0;
>> -
>> -	/* AEQ INT */
>> -	aemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> -	roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> -		     masken);
>> -	roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AE_IRQ_MASK_S, masken);
>> -	roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, aemask_val);
>> -
>> -	/* CEQ INT */
>> -	for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
>> -		/* IRQ mask */
>> -		roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> -			   i * CEQ_REG_OFFSET, masken);
>> -	}
>> -}
>> -
>> -static void hns_roce_ce_int_default_cfg(struct hns_roce_dev *hr_dev)
>> -{
>> -	/* Configure ce int interval */
>> -	roce_write(hr_dev, ROCEE_CAEP_CE_INTERVAL_CFG_REG,
>> -		   HNS_ROCE_CEQ_DEFAULT_INTERVAL);
>> -
>> -	/* Configure ce int burst num */
>> -	roce_write(hr_dev, ROCEE_CAEP_CE_BURST_NUM_CFG_REG,
>> -		   HNS_ROCE_CEQ_DEFAULT_BURST_NUM);
>> -}
>> -
>> -int hns_roce_init_eq_table(struct hns_roce_dev *hr_dev)
>> -{
>> -	struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
>> -	struct device *dev = &hr_dev->pdev->dev;
>> -	struct hns_roce_eq *eq = NULL;
>> -	int eq_num = 0;
>> -	int ret = 0;
>> -	int i = 0;
>> -	int j = 0;
>> -
>> -	eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
>> -	eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL);
>> -	if (!eq_table->eq)
>> -		return -ENOMEM;
>> -
>> -	eq_table->eqc_base = kcalloc(eq_num, sizeof(*eq_table->eqc_base),
>> -				     GFP_KERNEL);
>> -	if (!eq_table->eqc_base) {
>> -		ret = -ENOMEM;
>> -		goto err_eqc_base_alloc_fail;
>> -	}
>> -
>> -	for (i = 0; i < eq_num; i++) {
>> -		eq = &eq_table->eq[i];
>> -		eq->hr_dev = hr_dev;
>> -		eq->eqn = i;
>> -		eq->irq = hr_dev->irq[i];
>> -		eq->log_page_size = PAGE_SHIFT;
>> -
>> -		if (i < hr_dev->caps.num_comp_vectors) {
>> -			/* CEQ */
>> -			eq_table->eqc_base[i] = hr_dev->reg_base +
>> -						ROCEE_CAEP_CEQC_SHIFT_0_REG +
>> -						HNS_ROCE_CEQC_REG_OFFSET * i;
>> -			eq->type_flag = HNS_ROCE_CEQ;
>> -			eq->doorbell = hr_dev->reg_base +
>> -				       ROCEE_CAEP_CEQC_CONS_IDX_0_REG +
>> -				       HNS_ROCE_CEQC_REG_OFFSET * i;
>> -			eq->entries = hr_dev->caps.ceqe_depth[i];
>> -			eq->log_entries = ilog2(eq->entries);
>> -			eq->eqe_size = sizeof(struct hns_roce_ceqe);
>> -		} else {
>> -			/* AEQ */
>> -			eq_table->eqc_base[i] = hr_dev->reg_base +
>> -						ROCEE_CAEP_AEQC_AEQE_SHIFT_REG;
>> -			eq->type_flag = HNS_ROCE_AEQ;
>> -			eq->doorbell = hr_dev->reg_base +
>> -				       ROCEE_CAEP_AEQE_CONS_IDX_REG;
>> -			eq->entries = hr_dev->caps.aeqe_depth;
>> -			eq->log_entries = ilog2(eq->entries);
>> -			eq->eqe_size = sizeof(struct hns_roce_aeqe);
>> -		}
>> -	}
>> -
>> -	/* Disable irq */
>> -	hns_roce_int_mask_en(hr_dev);
>> -
>> -	/* Configure CE irq interval and burst num */
>> -	hns_roce_ce_int_default_cfg(hr_dev);
>> -
>> -	for (i = 0; i < eq_num; i++) {
>> -		ret = hns_roce_create_eq(hr_dev, &eq_table->eq[i]);
>> -		if (ret) {
>> -			dev_err(dev, "eq create failed\n");
>> -			goto err_create_eq_fail;
>> -		}
>> -	}
>> -
>> -	for (j = 0; j < eq_num; j++) {
>> -		ret = request_irq(eq_table->eq[j].irq, hns_roce_msi_x_interrupt,
>> -				  0, hr_dev->irq_names[j], eq_table->eq + j);
>> -		if (ret) {
>> -			dev_err(dev, "request irq error!\n");
>> -			goto err_request_irq_fail;
>> -		}
>> -	}
>> -
>> -	for (i = 0; i < eq_num; i++)
>> -		hns_roce_enable_eq(hr_dev, i, EQ_ENABLE);
>> -
>> -	return 0;
>> -
>> -err_request_irq_fail:
>> -	for (j = j - 1; j >= 0; j--)
>> -		free_irq(eq_table->eq[j].irq, eq_table->eq + j);
>> -
>> -err_create_eq_fail:
>> -	for (i = i - 1; i >= 0; i--)
>> -		hns_roce_free_eq(hr_dev, &eq_table->eq[i]);
>> -
>> -	kfree(eq_table->eqc_base);
>> -
>> -err_eqc_base_alloc_fail:
>> -	kfree(eq_table->eq);
>> -
>> -	return ret;
>> -}
>> -
>> -void hns_roce_cleanup_eq_table(struct hns_roce_dev *hr_dev)
>> -{
>> -	int i;
>> -	int eq_num;
>> -	struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
>> -
>> -	eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
>> -	for (i = 0; i < eq_num; i++) {
>> -		/* Disable EQ */
>> -		hns_roce_enable_eq(hr_dev, i, EQ_DISABLE);
>> -
>> -		free_irq(eq_table->eq[i].irq, eq_table->eq + i);
>> -
>> -		hns_roce_free_eq(hr_dev, &eq_table->eq[i]);
>> -	}
>> -
>> -	kfree(eq_table->eqc_base);
>> -	kfree(eq_table->eq);
>> -}
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_eq.h b/drivers/infiniband/hw/hns/hns_roce_eq.h
>> deleted file mode 100644
>> index c6d212d..0000000
>> --- a/drivers/infiniband/hw/hns/hns_roce_eq.h
>> +++ /dev/null
>> @@ -1,134 +0,0 @@
>> -/*
>> - * Copyright (c) 2016 Hisilicon Limited.
>> - *
>> - * This software is available to you under a choice of one of two
>> - * licenses.  You may choose to be licensed under the terms of the GNU
>> - * General Public License (GPL) Version 2, available from the file
>> - * COPYING in the main directory of this source tree, or the
>> - * OpenIB.org BSD license below:
>> - *
>> - *     Redistribution and use in source and binary forms, with or
>> - *     without modification, are permitted provided that the following
>> - *     conditions are met:
>> - *
>> - *      - Redistributions of source code must retain the above
>> - *        copyright notice, this list of conditions and the following
>> - *        disclaimer.
>> - *
>> - *      - Redistributions in binary form must reproduce the above
>> - *        copyright notice, this list of conditions and the following
>> - *        disclaimer in the documentation and/or other materials
>> - *        provided with the distribution.
>> - *
>> - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
>> - * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
>> - * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
>> - * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
>> - * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
>> - * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
>> - * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
>> - * SOFTWARE.
>> - */
>> -
>> -#ifndef _HNS_ROCE_EQ_H
>> -#define _HNS_ROCE_EQ_H
>> -
>> -#define HNS_ROCE_CEQ			1
>> -#define HNS_ROCE_AEQ			2
>> -
>> -#define HNS_ROCE_CEQ_ENTRY_SIZE		0x4
>> -#define HNS_ROCE_AEQ_ENTRY_SIZE		0x10
>> -#define HNS_ROCE_CEQC_REG_OFFSET	0x18
>> -
>> -#define HNS_ROCE_CEQ_DEFAULT_INTERVAL	0x10
>> -#define HNS_ROCE_CEQ_DEFAULT_BURST_NUM	0x10
>> -
>> -#define HNS_ROCE_INT_MASK_DISABLE	0
>> -#define HNS_ROCE_INT_MASK_ENABLE	1
>> -
>> -#define EQ_ENABLE			1
>> -#define EQ_DISABLE			0
>> -#define CONS_INDEX_MASK			0xffff
>> -
>> -#define CEQ_REG_OFFSET			0x18
>> -
>> -enum {
>> -	HNS_ROCE_EQ_STAT_INVALID  = 0,
>> -	HNS_ROCE_EQ_STAT_VALID    = 2,
>> -};
>> -
>> -struct hns_roce_aeqe {
>> -	u32 asyn;
>> -	union {
>> -		struct {
>> -			u32 qp;
>> -			u32 rsv0;
>> -			u32 rsv1;
>> -		} qp_event;
>> -
>> -		struct {
>> -			u32 cq;
>> -			u32 rsv0;
>> -			u32 rsv1;
>> -		} cq_event;
>> -
>> -		struct {
>> -			u32 port;
>> -			u32 rsv0;
>> -			u32 rsv1;
>> -		} port_event;
>> -
>> -		struct {
>> -			u32 ceqe;
>> -			u32 rsv0;
>> -			u32 rsv1;
>> -		} ce_event;
>> -
>> -		struct {
>> -			__le64  out_param;
>> -			__le16  token;
>> -			u8	status;
>> -			u8	rsv0;
>> -		} __packed cmd;
>> -	 } event;
>> -};
>> -
>> -#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S 16
>> -#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M   \
>> -	(((1UL << 8) - 1) << HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S)
>> -
>> -#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S 24
>> -#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M   \
>> -	(((1UL << 7) - 1) << HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)
>> -
>> -#define HNS_ROCE_AEQE_U32_4_OWNER_S 31
>> -
>> -#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S 0
>> -#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M   \
>> -	(((1UL << 24) - 1) << HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S)
>> -
>> -#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S 25
>> -#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M   \
>> -	(((1UL << 3) - 1) << HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S)
>> -
>> -#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S 0
>> -#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M   \
>> -	(((1UL << 16) - 1) << HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S)
>> -
>> -#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S 0
>> -#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M   \
>> -	(((1UL << 5) - 1) << HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S)
>> -
>> -struct hns_roce_ceqe {
>> -	union {
>> -		int		comp;
>> -	} ceqe;
>> -};
>> -
>> -#define HNS_ROCE_CEQE_CEQE_COMP_OWNER_S	0
>> -
>> -#define HNS_ROCE_CEQE_CEQE_COMP_CQN_S 16
>> -#define HNS_ROCE_CEQE_CEQE_COMP_CQN_M   \
>> -	(((1UL << 16) - 1) << HNS_ROCE_CEQE_CEQE_COMP_CQN_S)
>> -
>> -#endif /* _HNS_ROCE_EQ_H */
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v1.c b/drivers/infiniband/hw/hns/hns_roce_hw_v1.c
>> index af27168..6100ace 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_hw_v1.c
>> +++ b/drivers/infiniband/hw/hns/hns_roce_hw_v1.c
>> @@ -33,6 +33,7 @@
>>  #include <linux/platform_device.h>
>>  #include <linux/acpi.h>
>>  #include <linux/etherdevice.h>
>> +#include <linux/interrupt.h>
>>  #include <linux/of.h>
>>  #include <linux/of_platform.h>
>>  #include <rdma/ib_umem.h>
>> @@ -1492,9 +1493,9 @@ static int hns_roce_v1_profile(struct hns_roce_dev *hr_dev)
>>  	caps->max_sq_inline	= HNS_ROCE_V1_INLINE_SIZE;
>>  	caps->num_uars		= HNS_ROCE_V1_UAR_NUM;
>>  	caps->phy_num_uars	= HNS_ROCE_V1_PHY_UAR_NUM;
>> -	caps->num_aeq_vectors	= HNS_ROCE_AEQE_VEC_NUM;
>> -	caps->num_comp_vectors	= HNS_ROCE_COMP_VEC_NUM;
>> -	caps->num_other_vectors	= HNS_ROCE_AEQE_OF_VEC_NUM;
>> +	caps->num_aeq_vectors	= HNS_ROCE_V1_AEQE_VEC_NUM;
>> +	caps->num_comp_vectors	= HNS_ROCE_V1_COMP_VEC_NUM;
>> +	caps->num_other_vectors	= HNS_ROCE_V1_ABNORMAL_VEC_NUM;
>>  	caps->num_mtpts		= HNS_ROCE_V1_MAX_MTPT_NUM;
>>  	caps->num_mtt_segs	= HNS_ROCE_V1_MAX_MTT_SEGS;
>>  	caps->num_pds		= HNS_ROCE_V1_MAX_PD_NUM;
>> @@ -1529,10 +1530,8 @@ static int hns_roce_v1_profile(struct hns_roce_dev *hr_dev)
>>  						 caps->num_ports + 1;
>>  	}
>>  
>> -	for (i = 0; i < caps->num_comp_vectors; i++)
>> -		caps->ceqe_depth[i] = HNS_ROCE_V1_NUM_COMP_EQE;
>> -
>> -	caps->aeqe_depth = HNS_ROCE_V1_NUM_ASYNC_EQE;
>> +	caps->ceqe_depth = HNS_ROCE_V1_COMP_EQE_NUM;
>> +	caps->aeqe_depth = HNS_ROCE_V1_ASYNC_EQE_NUM;
>>  	caps->local_ca_ack_delay = le32_to_cpu(roce_read(hr_dev,
>>  							 ROCEE_ACK_DELAY_REG));
>>  	caps->max_mtu = IB_MTU_2048;
>> @@ -3960,6 +3959,727 @@ static int hns_roce_v1_destroy_cq(struct ib_cq *ibcq)
>>  	return ret;
>>  }
>>  
>> +static void set_eq_cons_index_v1(struct hns_roce_eq *eq, int req_not)
>> +{
>> +	roce_raw_write((eq->cons_index & HNS_ROCE_V1_CONS_IDX_M) |
>> +		      (req_not << eq->log_entries), eq->doorbell);
>> +	/* Memory barrier */
>> +	mb();
>> +}
>> +
>> +static void hns_roce_v1_wq_catas_err_handle(struct hns_roce_dev *hr_dev,
>> +					    struct hns_roce_aeqe *aeqe, int qpn)
>> +{
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +
>> +	dev_warn(dev, "Local Work Queue Catastrophic Error.\n");
>> +	switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> +			       HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> +	case HNS_ROCE_LWQCE_QPC_ERROR:
>> +		dev_warn(dev, "QP %d, QPC error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LWQCE_MTU_ERROR:
>> +		dev_warn(dev, "QP %d, MTU error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LWQCE_WQE_BA_ADDR_ERROR:
>> +		dev_warn(dev, "QP %d, WQE BA addr error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LWQCE_WQE_ADDR_ERROR:
>> +		dev_warn(dev, "QP %d, WQE addr error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LWQCE_SQ_WQE_SHIFT_ERROR:
>> +		dev_warn(dev, "QP %d, WQE shift error\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LWQCE_SL_ERROR:
>> +		dev_warn(dev, "QP %d, SL error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LWQCE_PORT_ERROR:
>> +		dev_warn(dev, "QP %d, port error.\n", qpn);
>> +		break;
>> +	default:
>> +		break;
>> +	}
>> +}
>> +
>> +static void hns_roce_v1_local_wq_access_err_handle(struct hns_roce_dev *hr_dev,
>> +						   struct hns_roce_aeqe *aeqe,
>> +						   int qpn)
>> +{
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +
>> +	dev_warn(dev, "Local Access Violation Work Queue Error.\n");
>> +	switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> +			       HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> +	case HNS_ROCE_LAVWQE_R_KEY_VIOLATION:
>> +		dev_warn(dev, "QP %d, R_key violation.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LAVWQE_LENGTH_ERROR:
>> +		dev_warn(dev, "QP %d, length error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LAVWQE_VA_ERROR:
>> +		dev_warn(dev, "QP %d, VA error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LAVWQE_PD_ERROR:
>> +		dev_err(dev, "QP %d, PD error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LAVWQE_RW_ACC_ERROR:
>> +		dev_warn(dev, "QP %d, rw acc error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LAVWQE_KEY_STATE_ERROR:
>> +		dev_warn(dev, "QP %d, key state error.\n", qpn);
>> +		break;
>> +	case HNS_ROCE_LAVWQE_MR_OPERATION_ERROR:
>> +		dev_warn(dev, "QP %d, MR operation error.\n", qpn);
>> +		break;
>> +	default:
>> +		break;
>> +	}
>> +}
>> +
>> +static void hns_roce_v1_qp_err_handle(struct hns_roce_dev *hr_dev,
>> +				      struct hns_roce_aeqe *aeqe,
>> +				      int event_type)
>> +{
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +	int phy_port;
>> +	int qpn;
>> +
>> +	qpn = roce_get_field(aeqe->event.qp_event.qp,
>> +			     HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M,
>> +			     HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S);
>> +	phy_port = roce_get_field(aeqe->event.qp_event.qp,
>> +				  HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M,
>> +				  HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S);
>> +	if (qpn <= 1)
>> +		qpn = HNS_ROCE_MAX_PORTS * qpn + phy_port;
>> +
>> +	switch (event_type) {
>> +	case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
>> +		dev_warn(dev, "Invalid Req Local Work Queue Error.\n"
>> +			 "QP %d, phy_port %d.\n", qpn, phy_port);
>> +		break;
>> +	case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
>> +		hns_roce_v1_wq_catas_err_handle(hr_dev, aeqe, qpn);
>> +		break;
>> +	case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
>> +		hns_roce_v1_local_wq_access_err_handle(hr_dev, aeqe, qpn);
>> +		break;
>> +	default:
>> +		break;
>> +	}
>> +
>> +	hns_roce_qp_event(hr_dev, qpn, event_type);
>> +}
>> +
>> +static void hns_roce_v1_cq_err_handle(struct hns_roce_dev *hr_dev,
>> +				      struct hns_roce_aeqe *aeqe,
>> +				      int event_type)
>> +{
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +	u32 cqn;
>> +
>> +	cqn = le32_to_cpu(roce_get_field(aeqe->event.cq_event.cq,
>> +			  HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M,
>> +			  HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S));
>> +
>> +	switch (event_type) {
>> +	case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
>> +		dev_warn(dev, "CQ 0x%x access err.\n", cqn);
>> +		break;
>> +	case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
>> +		dev_warn(dev, "CQ 0x%x overflow\n", cqn);
>> +		break;
>> +	case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
>> +		dev_warn(dev, "CQ 0x%x ID invalid.\n", cqn);
>> +		break;
>> +	default:
>> +		break;
>> +	}
>> +
>> +	hns_roce_cq_event(hr_dev, cqn, event_type);
>> +}
>> +
>> +static void hns_roce_v1_db_overflow_handle(struct hns_roce_dev *hr_dev,
>> +					   struct hns_roce_aeqe *aeqe)
>> +{
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +
>> +	switch (roce_get_field(aeqe->asyn, HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M,
>> +			       HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S)) {
>> +	case HNS_ROCE_DB_SUBTYPE_SDB_OVF:
>> +		dev_warn(dev, "SDB overflow.\n");
>> +		break;
>> +	case HNS_ROCE_DB_SUBTYPE_SDB_ALM_OVF:
>> +		dev_warn(dev, "SDB almost overflow.\n");
>> +		break;
>> +	case HNS_ROCE_DB_SUBTYPE_SDB_ALM_EMP:
>> +		dev_warn(dev, "SDB almost empty.\n");
>> +		break;
>> +	case HNS_ROCE_DB_SUBTYPE_ODB_OVF:
>> +		dev_warn(dev, "ODB overflow.\n");
>> +		break;
>> +	case HNS_ROCE_DB_SUBTYPE_ODB_ALM_OVF:
>> +		dev_warn(dev, "ODB almost overflow.\n");
>> +		break;
>> +	case HNS_ROCE_DB_SUBTYPE_ODB_ALM_EMP:
>> +		dev_warn(dev, "SDB almost empty.\n");
>> +		break;
>> +	default:
>> +		break;
>> +	}
>> +}
>> +
>> +static struct hns_roce_aeqe *get_aeqe_v1(struct hns_roce_eq *eq, u32 entry)
>> +{
>> +	unsigned long off = (entry & (eq->entries - 1)) *
>> +			     HNS_ROCE_AEQ_ENTRY_SIZE;
>> +
>> +	return (struct hns_roce_aeqe *)((u8 *)
>> +		(eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
>> +		off % HNS_ROCE_BA_SIZE);
>> +}
>> +
>> +static struct hns_roce_aeqe *next_aeqe_sw_v1(struct hns_roce_eq *eq)
>> +{
>> +	struct hns_roce_aeqe *aeqe = get_aeqe_v1(eq, eq->cons_index);
>> +
>> +	return (roce_get_bit(aeqe->asyn, HNS_ROCE_AEQE_U32_4_OWNER_S) ^
>> +		!!(eq->cons_index & eq->entries)) ? aeqe : NULL;
>> +}
>> +
>> +static int hns_roce_v1_aeq_int(struct hns_roce_dev *hr_dev,
>> +			       struct hns_roce_eq *eq)
>> +{
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +	struct hns_roce_aeqe *aeqe;
>> +	int aeqes_found = 0;
>> +	int event_type;
>> +
>> +	while ((aeqe = next_aeqe_sw_v1(eq))) {
>> +		dev_dbg(dev, "aeqe = %p, aeqe->asyn.event_type = 0x%lx\n", aeqe,
>> +			roce_get_field(aeqe->asyn,
>> +				       HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> +				       HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S));
>> +		/* Memory barrier */
>> +		rmb();
>> +
>> +		event_type = roce_get_field(aeqe->asyn,
>> +					    HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M,
>> +					    HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S);
>> +		switch (event_type) {
>> +		case HNS_ROCE_EVENT_TYPE_PATH_MIG:
>> +			dev_warn(dev, "PATH MIG not supported\n");
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_COMM_EST:
>> +			dev_warn(dev, "COMMUNICATION established\n");
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_SQ_DRAINED:
>> +			dev_warn(dev, "SQ DRAINED not supported\n");
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_PATH_MIG_FAILED:
>> +			dev_warn(dev, "PATH MIG failed\n");
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_INV_REQ_LOCAL_WQ_ERROR:
>> +		case HNS_ROCE_EVENT_TYPE_WQ_CATAS_ERROR:
>> +		case HNS_ROCE_EVENT_TYPE_LOCAL_WQ_ACCESS_ERROR:
>> +			hns_roce_v1_qp_err_handle(hr_dev, aeqe, event_type);
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_SRQ_LIMIT_REACH:
>> +		case HNS_ROCE_EVENT_TYPE_SRQ_CATAS_ERROR:
>> +		case HNS_ROCE_EVENT_TYPE_SRQ_LAST_WQE_REACH:
>> +			dev_warn(dev, "SRQ not support!\n");
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_CQ_ACCESS_ERROR:
>> +		case HNS_ROCE_EVENT_TYPE_CQ_OVERFLOW:
>> +		case HNS_ROCE_EVENT_TYPE_CQ_ID_INVALID:
>> +			hns_roce_v1_cq_err_handle(hr_dev, aeqe, event_type);
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_PORT_CHANGE:
>> +			dev_warn(dev, "port change.\n");
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_MB:
>> +			hns_roce_cmd_event(hr_dev,
>> +					   le16_to_cpu(aeqe->event.cmd.token),
>> +					   aeqe->event.cmd.status,
>> +					   le64_to_cpu(aeqe->event.cmd.out_param
>> +					   ));
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_DB_OVERFLOW:
>> +			hns_roce_v1_db_overflow_handle(hr_dev, aeqe);
>> +			break;
>> +		case HNS_ROCE_EVENT_TYPE_CEQ_OVERFLOW:
>> +			dev_warn(dev, "CEQ 0x%lx overflow.\n",
>> +			roce_get_field(aeqe->event.ce_event.ceqe,
>> +				     HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M,
>> +				     HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S));
>> +			break;
>> +		default:
>> +			dev_warn(dev, "Unhandled event %d on EQ %d at idx %u.\n",
>> +				 event_type, eq->eqn, eq->cons_index);
>> +			break;
>> +		}
>> +
>> +		eq->cons_index++;
>> +		aeqes_found = 1;
>> +
>> +		if (eq->cons_index > 2 * hr_dev->caps.aeqe_depth - 1) {
>> +			dev_warn(dev, "cons_index overflow, set back to 0.\n");
>> +			eq->cons_index = 0;
>> +		}
>> +	}
>> +
>> +	set_eq_cons_index_v1(eq, 0);
>> +
>> +	return aeqes_found;
>> +}
>> +
>> +static struct hns_roce_ceqe *get_ceqe_v1(struct hns_roce_eq *eq, u32 entry)
>> +{
>> +	unsigned long off = (entry & (eq->entries - 1)) *
>> +			     HNS_ROCE_CEQ_ENTRY_SIZE;
>> +
>> +	return (struct hns_roce_ceqe *)((u8 *)
>> +			(eq->buf_list[off / HNS_ROCE_BA_SIZE].buf) +
>> +			off % HNS_ROCE_BA_SIZE);
>> +}
>> +
>> +static struct hns_roce_ceqe *next_ceqe_sw_v1(struct hns_roce_eq *eq)
>> +{
>> +	struct hns_roce_ceqe *ceqe = get_ceqe_v1(eq, eq->cons_index);
>> +
>> +	return (!!(roce_get_bit(ceqe->comp,
>> +		HNS_ROCE_CEQE_CEQE_COMP_OWNER_S))) ^
>> +		(!!(eq->cons_index & eq->entries)) ? ceqe : NULL;
>> +}
>> +
>> +static int hns_roce_v1_ceq_int(struct hns_roce_dev *hr_dev,
>> +			       struct hns_roce_eq *eq)
>> +{
>> +	struct hns_roce_ceqe *ceqe;
>> +	int ceqes_found = 0;
>> +	u32 cqn;
>> +
>> +	while ((ceqe = next_ceqe_sw_v1(eq))) {
>> +		/* Memory barrier */
>> +		rmb();
>> +		cqn = roce_get_field(ceqe->comp,
>> +				     HNS_ROCE_CEQE_CEQE_COMP_CQN_M,
>> +				     HNS_ROCE_CEQE_CEQE_COMP_CQN_S);
>> +		hns_roce_cq_completion(hr_dev, cqn);
>> +
>> +		++eq->cons_index;
>> +		ceqes_found = 1;
>> +
>> +		if (eq->cons_index > 2 * hr_dev->caps.ceqe_depth - 1) {
>> +			dev_warn(&eq->hr_dev->pdev->dev,
>> +				"cons_index overflow, set back to 0.\n");
>> +			eq->cons_index = 0;
>> +		}
>> +	}
>> +
>> +	set_eq_cons_index_v1(eq, 0);
>> +
>> +	return ceqes_found;
>> +}
>> +
>> +static irqreturn_t hns_roce_v1_msix_interrupt_eq(int irq, void *eq_ptr)
>> +{
>> +	struct hns_roce_eq  *eq  = eq_ptr;
>> +	struct hns_roce_dev *hr_dev = eq->hr_dev;
>> +	int int_work = 0;
>> +
>> +	if (eq->type_flag == HNS_ROCE_CEQ)
>> +		/* CEQ irq routine, CEQ is pulse irq, not clear */
>> +		int_work = hns_roce_v1_ceq_int(hr_dev, eq);
>> +	else
>> +		/* AEQ irq routine, AEQ is pulse irq, not clear */
>> +		int_work = hns_roce_v1_aeq_int(hr_dev, eq);
>> +
>> +	return IRQ_RETVAL(int_work);
>> +}
>> +
>> +static irqreturn_t hns_roce_v1_msix_interrupt_abn(int irq, void *dev_id)
>> +{
>> +	struct hns_roce_dev *hr_dev = dev_id;
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +	int int_work = 0;
>> +	u32 caepaemask_val;
>> +	u32 cealmovf_val;
>> +	u32 caepaest_val;
>> +	u32 aeshift_val;
>> +	u32 ceshift_val;
>> +	u32 cemask_val;
>> +	int i;
>> +
>> +	/*
>> +	 * Abnormal interrupt:
>> +	 * AEQ overflow, ECC multi-bit err, CEQ overflow must clear
>> +	 * interrupt, mask irq, clear irq, cancel mask operation
>> +	 */
>> +	aeshift_val = roce_read(hr_dev, ROCEE_CAEP_AEQC_AEQE_SHIFT_REG);
>> +
>> +	/* AEQE overflow */
>> +	if (roce_get_bit(aeshift_val,
>> +		ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQ_ALM_OVF_INT_ST_S) == 1) {
>> +		dev_warn(dev, "AEQ overflow!\n");
>> +
>> +		/* Set mask */
>> +		caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> +		roce_set_bit(caepaemask_val,
>> +			     ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> +			     HNS_ROCE_INT_MASK_ENABLE);
>> +		roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
>> +
>> +		/* Clear int state(INT_WC : write 1 clear) */
>> +		caepaest_val = roce_read(hr_dev, ROCEE_CAEP_AE_ST_REG);
>> +		roce_set_bit(caepaest_val,
>> +			     ROCEE_CAEP_AE_ST_CAEP_AEQ_ALM_OVF_S, 1);
>> +		roce_write(hr_dev, ROCEE_CAEP_AE_ST_REG, caepaest_val);
>> +
>> +		/* Clear mask */
>> +		caepaemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> +		roce_set_bit(caepaemask_val,
>> +			     ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> +			     HNS_ROCE_INT_MASK_DISABLE);
>> +		roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, caepaemask_val);
>> +	}
>> +
>> +	/* CEQ almost overflow */
>> +	for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
>> +		ceshift_val = roce_read(hr_dev, ROCEE_CAEP_CEQC_SHIFT_0_REG +
>> +					i * CEQ_REG_OFFSET);
>> +
>> +		if (roce_get_bit(ceshift_val,
>> +			ROCEE_CAEP_CEQC_SHIFT_CAEP_CEQ_ALM_OVF_INT_ST_S) == 1) {
>> +			dev_warn(dev, "CEQ[%d] almost overflow!\n", i);
>> +			int_work++;
>> +
>> +			/* Set mask */
>> +			cemask_val = roce_read(hr_dev,
>> +					       ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> +					       i * CEQ_REG_OFFSET);
>> +			roce_set_bit(cemask_val,
>> +				ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
>> +				HNS_ROCE_INT_MASK_ENABLE);
>> +			roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> +				   i * CEQ_REG_OFFSET, cemask_val);
>> +
>> +			/* Clear int state(INT_WC : write 1 clear) */
>> +			cealmovf_val = roce_read(hr_dev,
>> +				       ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
>> +				       i * CEQ_REG_OFFSET);
>> +			roce_set_bit(cealmovf_val,
>> +				     ROCEE_CAEP_CEQ_ALM_OVF_CAEP_CEQ_ALM_OVF_S,
>> +				     1);
>> +			roce_write(hr_dev, ROCEE_CAEP_CEQ_ALM_OVF_0_REG +
>> +				   i * CEQ_REG_OFFSET, cealmovf_val);
>> +
>> +			/* Clear mask */
>> +			cemask_val = roce_read(hr_dev,
>> +				     ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> +				     i * CEQ_REG_OFFSET);
>> +			roce_set_bit(cemask_val,
>> +			       ROCEE_CAEP_CE_IRQ_MASK_CAEP_CEQ_ALM_OVF_MASK_S,
>> +			       HNS_ROCE_INT_MASK_DISABLE);
>> +			roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> +				   i * CEQ_REG_OFFSET, cemask_val);
>> +		}
>> +	}
>> +
>> +	/* ECC multi-bit error alarm */
>> +	dev_warn(dev, "ECC UCERR ALARM: 0x%x, 0x%x, 0x%x\n",
>> +		 roce_read(hr_dev, ROCEE_ECC_UCERR_ALM0_REG),
>> +		 roce_read(hr_dev, ROCEE_ECC_UCERR_ALM1_REG),
>> +		 roce_read(hr_dev, ROCEE_ECC_UCERR_ALM2_REG));
>> +
>> +	dev_warn(dev, "ECC CERR ALARM: 0x%x, 0x%x, 0x%x\n",
>> +		 roce_read(hr_dev, ROCEE_ECC_CERR_ALM0_REG),
>> +		 roce_read(hr_dev, ROCEE_ECC_CERR_ALM1_REG),
>> +		 roce_read(hr_dev, ROCEE_ECC_CERR_ALM2_REG));
>> +
>> +	return IRQ_RETVAL(int_work);
>> +}
>> +
>> +static void hns_roce_v1_int_mask_enable(struct hns_roce_dev *hr_dev)
>> +{
>> +	u32 aemask_val;
>> +	int masken = 0;
>> +	int i;
>> +
>> +	/* AEQ INT */
>> +	aemask_val = roce_read(hr_dev, ROCEE_CAEP_AE_MASK_REG);
>> +	roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AEQ_ALM_OVF_MASK_S,
>> +		     masken);
>> +	roce_set_bit(aemask_val, ROCEE_CAEP_AE_MASK_CAEP_AE_IRQ_MASK_S, masken);
>> +	roce_write(hr_dev, ROCEE_CAEP_AE_MASK_REG, aemask_val);
>> +
>> +	/* CEQ INT */
>> +	for (i = 0; i < hr_dev->caps.num_comp_vectors; i++) {
>> +		/* IRQ mask */
>> +		roce_write(hr_dev, ROCEE_CAEP_CE_IRQ_MASK_0_REG +
>> +			   i * CEQ_REG_OFFSET, masken);
>> +	}
>> +}
>> +
>> +static void hns_roce_v1_free_eq(struct hns_roce_dev *hr_dev,
>> +				struct hns_roce_eq *eq)
>> +{
>> +	int npages = (PAGE_ALIGN(eq->eqe_size * eq->entries) +
>> +		      HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
>> +	int i;
>> +
>> +	if (!eq->buf_list)
>> +		return;
>> +
>> +	for (i = 0; i < npages; ++i)
>> +		dma_free_coherent(&hr_dev->pdev->dev, HNS_ROCE_BA_SIZE,
>> +				  eq->buf_list[i].buf, eq->buf_list[i].map);
>> +
>> +	kfree(eq->buf_list);
>> +}
>> +
>> +static void hns_roce_v1_enable_eq(struct hns_roce_dev *hr_dev, int eq_num,
>> +				  int enable_flag)
>> +{
>> +	void __iomem *eqc = hr_dev->eq_table.eqc_base[eq_num];
>> +	u32 val;
>> +
>> +	val = readl(eqc);
>> +
>> +	if (enable_flag)
>> +		roce_set_field(val,
>> +			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> +			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> +			       HNS_ROCE_EQ_STAT_VALID);
>> +	else
>> +		roce_set_field(val,
>> +			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> +			       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> +			       HNS_ROCE_EQ_STAT_INVALID);
>> +	writel(val, eqc);
>> +}
>> +
>> +static int hns_roce_v1_create_eq(struct hns_roce_dev *hr_dev,
>> +				 struct hns_roce_eq *eq)
>> +{
>> +	void __iomem *eqc = hr_dev->eq_table.eqc_base[eq->eqn];
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +	dma_addr_t tmp_dma_addr;
>> +	u32 eqconsindx_val = 0;
>> +	u32 eqcuridx_val = 0;
>> +	u32 eqshift_val = 0;
>> +	int num_bas;
>> +	int ret;
>> +	int i;
>> +
>> +	num_bas = (PAGE_ALIGN(eq->entries * eq->eqe_size) +
>> +		   HNS_ROCE_BA_SIZE - 1) / HNS_ROCE_BA_SIZE;
>> +
>> +	if ((eq->entries * eq->eqe_size) > HNS_ROCE_BA_SIZE) {
>> +		dev_err(dev, "[error]eq buf %d gt ba size(%d) need bas=%d\n",
>> +			(eq->entries * eq->eqe_size), HNS_ROCE_BA_SIZE,
>> +			num_bas);
>> +		return -EINVAL;
>> +	}
>> +
>> +	eq->buf_list = kcalloc(num_bas, sizeof(*eq->buf_list), GFP_KERNEL);
>> +	if (!eq->buf_list)
>> +		return -ENOMEM;
>> +
>> +	for (i = 0; i < num_bas; ++i) {
>> +		eq->buf_list[i].buf = dma_alloc_coherent(dev, HNS_ROCE_BA_SIZE,
>> +							 &tmp_dma_addr,
>> +							 GFP_KERNEL);
>> +		if (!eq->buf_list[i].buf) {
>> +			ret = -ENOMEM;
>> +			goto err_out_free_pages;
>> +		}
>> +
>> +		eq->buf_list[i].map = tmp_dma_addr;
>> +		memset(eq->buf_list[i].buf, 0, HNS_ROCE_BA_SIZE);
>> +	}
>> +	eq->cons_index = 0;
>> +	roce_set_field(eqshift_val,
>> +		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_M,
>> +		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_STATE_S,
>> +		       HNS_ROCE_EQ_STAT_INVALID);
>> +	roce_set_field(eqshift_val,
>> +		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_M,
>> +		       ROCEE_CAEP_AEQC_AEQE_SHIFT_CAEP_AEQC_AEQE_SHIFT_S,
>> +		       eq->log_entries);
>> +	writel(eqshift_val, eqc);
>> +
>> +	/* Configure eq extended address 12~44bit */
>> +	writel((u32)(eq->buf_list[0].map >> 12), eqc + 4);
>> +
>> +	/*
>> +	 * Configure eq extended address 45~49 bit.
>> +	 * 44 = 32 + 12, When evaluating addr to hardware, shift 12 because of
>> +	 * using 4K page, and shift more 32 because of
>> +	 * caculating the high 32 bit value evaluated to hardware.
>> +	 */
>> +	roce_set_field(eqcuridx_val, ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_M,
>> +		       ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQ_BT_H_S,
>> +		       eq->buf_list[0].map >> 44);
>> +	roce_set_field(eqcuridx_val,
>> +		       ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_M,
>> +		       ROCEE_CAEP_AEQE_CUR_IDX_CAEP_AEQE_CUR_IDX_S, 0);
>> +	writel(eqcuridx_val, eqc + 8);
>> +
>> +	/* Configure eq consumer index */
>> +	roce_set_field(eqconsindx_val,
>> +		       ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_M,
>> +		       ROCEE_CAEP_AEQE_CONS_IDX_CAEP_AEQE_CONS_IDX_S, 0);
>> +	writel(eqconsindx_val, eqc + 0xc);
>> +
>> +	return 0;
>> +
>> +err_out_free_pages:
>> +	for (i -= 1; i >= 0; i--)
>> +		dma_free_coherent(dev, HNS_ROCE_BA_SIZE, eq->buf_list[i].buf,
>> +				  eq->buf_list[i].map);
>> +
>> +	kfree(eq->buf_list);
>> +	return ret;
>> +}
>> +
>> +static int hns_roce_v1_init_eq_table(struct hns_roce_dev *hr_dev)
>> +{
>> +	struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
>> +	struct device *dev = &hr_dev->pdev->dev;
>> +	struct hns_roce_eq *eq;
>> +	int irq_num;
>> +	int eq_num;
>> +	int ret;
>> +	int i, j;
>> +
>> +	eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
>> +	irq_num = eq_num + hr_dev->caps.num_other_vectors;
>> +
>> +	eq_table->eq = kcalloc(eq_num, sizeof(*eq_table->eq), GFP_KERNEL);
>> +	if (!eq_table->eq)
>> +		return -ENOMEM;
>> +
>> +	eq_table->eqc_base = kcalloc(eq_num, sizeof(*eq_table->eqc_base),
>> +				     GFP_KERNEL);
>> +	if (!eq_table->eqc_base) {
>> +		ret = -ENOMEM;
>> +		goto err_eqc_base_alloc_fail;
>> +	}
>> +
>> +	for (i = 0; i < eq_num; i++) {
>> +		eq = &eq_table->eq[i];
>> +		eq->hr_dev = hr_dev;
>> +		eq->eqn = i;
>> +		eq->irq = hr_dev->irq[i];
>> +		eq->log_page_size = PAGE_SHIFT;
>> +
>> +		if (i < hr_dev->caps.num_comp_vectors) {
>> +			/* CEQ */
>> +			eq_table->eqc_base[i] = hr_dev->reg_base +
>> +						ROCEE_CAEP_CEQC_SHIFT_0_REG +
>> +						CEQ_REG_OFFSET * i;
>> +			eq->type_flag = HNS_ROCE_CEQ;
>> +			eq->doorbell = hr_dev->reg_base +
>> +				       ROCEE_CAEP_CEQC_CONS_IDX_0_REG +
>> +				       CEQ_REG_OFFSET * i;
>> +			eq->entries = hr_dev->caps.ceqe_depth;
>> +			eq->log_entries = ilog2(eq->entries);
>> +			eq->eqe_size = HNS_ROCE_CEQ_ENTRY_SIZE;
>> +		} else {
>> +			/* AEQ */
>> +			eq_table->eqc_base[i] = hr_dev->reg_base +
>> +						ROCEE_CAEP_AEQC_AEQE_SHIFT_REG;
>> +			eq->type_flag = HNS_ROCE_AEQ;
>> +			eq->doorbell = hr_dev->reg_base +
>> +				       ROCEE_CAEP_AEQE_CONS_IDX_REG;
>> +			eq->entries = hr_dev->caps.aeqe_depth;
>> +			eq->log_entries = ilog2(eq->entries);
>> +			eq->eqe_size = HNS_ROCE_AEQ_ENTRY_SIZE;
>> +		}
>> +	}
>> +
>> +	/* Disable irq */
>> +	hns_roce_v1_int_mask_enable(hr_dev);
>> +
>> +	/* Configure ce int interval */
>> +	roce_write(hr_dev, ROCEE_CAEP_CE_INTERVAL_CFG_REG,
>> +		   HNS_ROCE_CEQ_DEFAULT_INTERVAL);
>> +
>> +	/* Configure ce int burst num */
>> +	roce_write(hr_dev, ROCEE_CAEP_CE_BURST_NUM_CFG_REG,
>> +		   HNS_ROCE_CEQ_DEFAULT_BURST_NUM);
>> +
>> +	for (i = 0; i < eq_num; i++) {
>> +		ret = hns_roce_v1_create_eq(hr_dev, &eq_table->eq[i]);
>> +		if (ret) {
>> +			dev_err(dev, "eq create failed\n");
>> +			goto err_create_eq_fail;
>> +		}
>> +	}
>> +
>> +	for (j = 0; j < irq_num; j++) {
>> +		if (j < eq_num)
>> +			ret = request_irq(hr_dev->irq[j],
>> +					  hns_roce_v1_msix_interrupt_eq, 0,
>> +					  hr_dev->irq_names[j],
>> +					  &eq_table->eq[j]);
>> +		else
>> +			ret = request_irq(hr_dev->irq[j],
>> +					  hns_roce_v1_msix_interrupt_abn, 0,
>> +					  hr_dev->irq_names[j], hr_dev);
>> +
>> +		if (ret) {
>> +			dev_err(dev, "request irq error!\n");
>> +			goto err_request_irq_fail;
>> +		}
>> +	}
>> +
>> +	for (i = 0; i < eq_num; i++)
>> +		hns_roce_v1_enable_eq(hr_dev, i, EQ_ENABLE);
>> +
>> +	return 0;
>> +
>> +err_request_irq_fail:
>> +	for (j -= 1; j >= 0; j--)
>> +		free_irq(hr_dev->irq[j], &eq_table->eq[j]);
>> +
>> +err_create_eq_fail:
>> +	for (i -= 1; i >= 0; i--)
>> +		hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]);
>> +
>> +	kfree(eq_table->eqc_base);
>> +
>> +err_eqc_base_alloc_fail:
>> +	kfree(eq_table->eq);
>> +
>> +	return ret;
>> +}
>> +
>> +static void hns_roce_v1_cleanup_eq_table(struct hns_roce_dev *hr_dev)
>> +{
>> +	struct hns_roce_eq_table *eq_table = &hr_dev->eq_table;
>> +	int irq_num;
>> +	int eq_num;
>> +	int i;
>> +
>> +	eq_num = hr_dev->caps.num_comp_vectors + hr_dev->caps.num_aeq_vectors;
>> +	irq_num = eq_num + hr_dev->caps.num_other_vectors;
>> +	for (i = 0; i < eq_num; i++) {
>> +		/* Disable EQ */
>> +		hns_roce_v1_enable_eq(hr_dev, i, EQ_DISABLE);
>> +
>> +		free_irq(hr_dev->irq[i], &eq_table->eq[i]);
>> +
>> +		hns_roce_v1_free_eq(hr_dev, &eq_table->eq[i]);
>> +	}
>> +	for (i = eq_num; i < irq_num; i++)
>> +		free_irq(hr_dev->irq[i], hr_dev);
>> +
>> +	kfree(eq_table->eqc_base);
>> +	kfree(eq_table->eq);
>> +}
>> +
>>  static const struct hns_roce_hw hns_roce_hw_v1 = {
>>  	.reset = hns_roce_v1_reset,
>>  	.hw_profile = hns_roce_v1_profile,
>> @@ -3983,6 +4703,8 @@ static int hns_roce_v1_destroy_cq(struct ib_cq *ibcq)
>>  	.poll_cq = hns_roce_v1_poll_cq,
>>  	.dereg_mr = hns_roce_v1_dereg_mr,
>>  	.destroy_cq = hns_roce_v1_destroy_cq,
>> +	.init_eq = hns_roce_v1_init_eq_table,
>> +	.cleanup_eq = hns_roce_v1_cleanup_eq_table,
>>  };
>>  
>>  static const struct of_device_id hns_roce_of_match[] = {
>> @@ -4132,14 +4854,14 @@ static int hns_roce_get_cfg(struct hns_roce_dev *hr_dev)
>>  	/* read the interrupt names from the DT or ACPI */
>>  	ret = device_property_read_string_array(dev, "interrupt-names",
>>  						hr_dev->irq_names,
>> -						HNS_ROCE_MAX_IRQ_NUM);
>> +						HNS_ROCE_V1_MAX_IRQ_NUM);
>>  	if (ret < 0) {
>>  		dev_err(dev, "couldn't get interrupt names from DT or ACPI!\n");
>>  		return ret;
>>  	}
>>  
>>  	/* fetch the interrupt numbers */
>> -	for (i = 0; i < HNS_ROCE_MAX_IRQ_NUM; i++) {
>> +	for (i = 0; i < HNS_ROCE_V1_MAX_IRQ_NUM; i++) {
>>  		hr_dev->irq[i] = platform_get_irq(hr_dev->pdev, i);
>>  		if (hr_dev->irq[i] <= 0) {
>>  			dev_err(dev, "platform get of irq[=%d] failed!\n", i);
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_hw_v1.h b/drivers/infiniband/hw/hns/hns_roce_hw_v1.h
>> index 21a07ef..b44ddd2 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_hw_v1.h
>> +++ b/drivers/infiniband/hw/hns/hns_roce_hw_v1.h
>> @@ -60,8 +60,13 @@
>>  #define HNS_ROCE_V1_GID_NUM				16
>>  #define HNS_ROCE_V1_RESV_QP				8
>>  
>> -#define HNS_ROCE_V1_NUM_COMP_EQE			0x8000
>> -#define HNS_ROCE_V1_NUM_ASYNC_EQE			0x400
>> +#define HNS_ROCE_V1_MAX_IRQ_NUM				34
>> +#define HNS_ROCE_V1_COMP_VEC_NUM			32
>> +#define HNS_ROCE_V1_AEQE_VEC_NUM			1
>> +#define HNS_ROCE_V1_ABNORMAL_VEC_NUM			1
>> +
>> +#define HNS_ROCE_V1_COMP_EQE_NUM			0x8000
>> +#define HNS_ROCE_V1_ASYNC_EQE_NUM			0x400
>>  
>>  #define HNS_ROCE_V1_QPC_ENTRY_SIZE			256
>>  #define HNS_ROCE_V1_IRRL_ENTRY_SIZE			8
>> @@ -159,6 +164,41 @@
>>  #define SDB_INV_CNT_OFFSET				8
>>  #define SDB_ST_CMP_VAL					8
>>  
>> +#define HNS_ROCE_CEQ_DEFAULT_INTERVAL			0x10
>> +#define HNS_ROCE_CEQ_DEFAULT_BURST_NUM			0x10
>> +
>> +#define HNS_ROCE_INT_MASK_DISABLE			0
>> +#define HNS_ROCE_INT_MASK_ENABLE			1
>> +
>> +#define CEQ_REG_OFFSET					0x18
>> +
>> +#define HNS_ROCE_CEQE_CEQE_COMP_OWNER_S	0
>> +
>> +#define HNS_ROCE_V1_CONS_IDX_M GENMASK(15, 0)
>> +
>> +#define HNS_ROCE_CEQE_CEQE_COMP_CQN_S 16
>> +#define HNS_ROCE_CEQE_CEQE_COMP_CQN_M GENMASK(31, 16)
>> +
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_S 16
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_TYPE_M GENMASK(23, 16)
>> +
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_S 24
>> +#define HNS_ROCE_AEQE_U32_4_EVENT_SUB_TYPE_M GENMASK(30, 24)
>> +
>> +#define HNS_ROCE_AEQE_U32_4_OWNER_S 31
>> +
>> +#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_S 0
>> +#define HNS_ROCE_AEQE_EVENT_QP_EVENT_QP_QPN_M GENMASK(23, 0)
>> +
>> +#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_S 25
>> +#define HNS_ROCE_AEQE_EVENT_QP_EVENT_PORT_NUM_M GENMASK(27, 25)
>> +
>> +#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_S 0
>> +#define HNS_ROCE_AEQE_EVENT_CQ_EVENT_CQ_CQN_M GENMASK(15, 0)
>> +
>> +#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_S 0
>> +#define HNS_ROCE_AEQE_EVENT_CE_EVENT_CEQE_CEQN_M GENMASK(4, 0)
>> +
>>  struct hns_roce_cq_context {
>>  	u32 cqc_byte_4;
>>  	u32 cq_bt_l;
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_main.c b/drivers/infiniband/hw/hns/hns_roce_main.c
>> index cf02ac2..aa0c242 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_main.c
>> +++ b/drivers/infiniband/hw/hns/hns_roce_main.c
>> @@ -748,12 +748,10 @@ int hns_roce_init(struct hns_roce_dev *hr_dev)
>>  		goto error_failed_cmd_init;
>>  	}
>>  
>> -	if (hr_dev->cmd_mod) {
>> -		ret = hns_roce_init_eq_table(hr_dev);
>> -		if (ret) {
>> -			dev_err(dev, "eq init failed!\n");
>> -			goto error_failed_eq_table;
>> -		}
>> +	ret = hr_dev->hw->init_eq(hr_dev);
>> +	if (ret) {
>> +		dev_err(dev, "eq init failed!\n");
>> +		goto error_failed_eq_table;
>>  	}
>>  
>>  	if (hr_dev->cmd_mod) {
>> @@ -805,8 +803,7 @@ int hns_roce_init(struct hns_roce_dev *hr_dev)
>>  		hns_roce_cmd_use_polling(hr_dev);
>>  
>>  error_failed_use_event:
>> -	if (hr_dev->cmd_mod)
>> -		hns_roce_cleanup_eq_table(hr_dev);
>> +	hr_dev->hw->cleanup_eq(hr_dev);
>>  
>>  error_failed_eq_table:
>>  	hns_roce_cmd_cleanup(hr_dev);
>> @@ -837,8 +834,7 @@ void hns_roce_exit(struct hns_roce_dev *hr_dev)
>>  	if (hr_dev->cmd_mod)
>>  		hns_roce_cmd_use_polling(hr_dev);
>>  
>> -	if (hr_dev->cmd_mod)
>> -		hns_roce_cleanup_eq_table(hr_dev);
>> +	hr_dev->hw->cleanup_eq(hr_dev);
>>  	hns_roce_cmd_cleanup(hr_dev);
>>  	if (hr_dev->hw->cmq_exit)
>>  		hr_dev->hw->cmq_exit(hr_dev);
>> diff --git a/drivers/infiniband/hw/hns/hns_roce_qp.c b/drivers/infiniband/hw/hns/hns_roce_qp.c
>> index 49586ec..69e2584 100644
>> --- a/drivers/infiniband/hw/hns/hns_roce_qp.c
>> +++ b/drivers/infiniband/hw/hns/hns_roce_qp.c
>> @@ -65,6 +65,7 @@ void hns_roce_qp_event(struct hns_roce_dev *hr_dev, u32 qpn, int event_type)
>>  	if (atomic_dec_and_test(&qp->refcount))
>>  		complete(&qp->free);
>>  }
>> +EXPORT_SYMBOL_GPL(hns_roce_qp_event);
>>  
>>  static void hns_roce_ib_qp_event(struct hns_roce_qp *hr_qp,
>>  				 enum hns_roce_event type)
>>
> 
> --
> To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> .
> 

--
To unsubscribe from this list: send the line "unsubscribe linux-rdma" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Photo]     [Yosemite News]     [Yosemite Photos]     [Linux Kernel]     [Linux SCSI]     [XFree86]
  Powered by Linux