Re: Re: Re: [PATCH for-next] RDMA/siw: fix SQ/RQ drain logic to support ib_drain_qp

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



-----"Krishnamraju Eraparaju" <krishna2@xxxxxxxxxxx> wrote: -----

>To: "Bernard Metzler" <BMT@xxxxxxxxxxxxxx>
>From: "Krishnamraju Eraparaju" <krishna2@xxxxxxxxxxx>
>Date: 10/01/2019 07:45PM
>Cc: "jgg@xxxxxxxx" <jgg@xxxxxxxx>, "linux-rdma@xxxxxxxxxxxxxxx"
><linux-rdma@xxxxxxxxxxxxxxx>, "Potnuri Bharat Teja"
><bharat@xxxxxxxxxxx>, "Nirranjan Kirubaharan" <nirranjan@xxxxxxxxxxx>
>Subject: [EXTERNAL] Re: Re: [PATCH for-next] RDMA/siw: fix SQ/RQ
>drain logic to support ib_drain_qp
>
>On Tuesday, October 10/01/19, 2019 at 15:56:45 +0000, Bernard Metzler
>wrote:
>> -----"Krishnamraju Eraparaju" <krishna2@xxxxxxxxxxx> wrote: -----
>> 
>> >To: "Bernard Metzler" <BMT@xxxxxxxxxxxxxx>
>> >From: "Krishnamraju Eraparaju" <krishna2@xxxxxxxxxxx>
>> >Date: 10/01/2019 11:52AM
>> >Cc: "jgg@xxxxxxxx" <jgg@xxxxxxxx>, "linux-rdma@xxxxxxxxxxxxxxx"
>> ><linux-rdma@xxxxxxxxxxxxxxx>, "Potnuri Bharat Teja"
>> ><bharat@xxxxxxxxxxx>, "Nirranjan Kirubaharan"
><nirranjan@xxxxxxxxxxx>
>> >Subject: [EXTERNAL] Re: [PATCH for-next] RDMA/siw: fix SQ/RQ drain
>> >logic to support ib_drain_qp
>> >
>> >On Monday, September 09/30/19, 2019 at 21:07:23 +0530, Bernard
>> >Metzler wrote:
>> >> -----"Krishnamraju Eraparaju" <krishna2@xxxxxxxxxxx> wrote:
>-----
>> >> 
>> >> >To: jgg@xxxxxxxx, bmt@xxxxxxxxxxxxxx
>> >> >From: "Krishnamraju Eraparaju" <krishna2@xxxxxxxxxxx>
>> >> >Date: 09/28/2019 12:16AM
>> >> >Cc: linux-rdma@xxxxxxxxxxxxxxx, bharat@xxxxxxxxxxx,
>> >> >nirranjan@xxxxxxxxxxx, "Krishnamraju Eraparaju"
>> >> ><krishna2@xxxxxxxxxxx>
>> >> >Subject: [EXTERNAL] [PATCH for-next] RDMA/siw: fix SQ/RQ drain
>> >logic
>> >> >to support ib_drain_qp
>> >> >
>> >> >The storage ULPs(iSER & NVMeOF) uses ib_drain_qp() to drain
>> >> >QP/CQ properly. But SIW is currently using it's own routines to
>> >> >drain SQ & RQ, which can't help ULPs to determine the last CQE.
>> >> >Failing to wait until last CQE causes touch after free issues:
>> >> 
>> >> Hi Krishna,
>> >> 
>> >> Before reviewing, please let me fully understand what is
>> >> going on here/why we need that patch.
>> >> 
>> >> Is this issue caused since the ulp expects the ib_drain_xx
>> >> driver method to be blocking until all completions are reaped,
>> >> and siw does not block?
>> >Yes, SIW is currently using provider-specific drain_qp logic,
>> >IE: siw_verbs_rq/sq_flush(), with this logic though all the SQ &
>RQ
>> >entries are flushed to CQ but ULPs cannot ensure when exactly the
>> >processing of all CQEs for those WRs, posted prior to
>ib_drain_xx(),
>> >got completed.
>> >Due to this uncertainity, sometimes iSER/NVMeOF driver(assuming
>all
>> >the CQEs are processed) proceed to release resouces/destroy_qp,
>> >causing touch after free issues.
>> >> 
>> >> Is the ULP expected to call ib_drain_xx only if no other
>> >> work is pending (SQ/RQ/CQ)? If not, shall all previous
>> >> work be completed with FLUSH error?
>> >In error cases(eg: link-down), I see iSER/NVMEOF drivers
>performing
>> >disconnect/drain_qp though there is some pending work to be
>> >processed.
>> >This may be valid due to the ERROR.
>> >And all that pending work gets completed with FLUSH error. 
>> 
>> OK understood.
>> 
>> Dropping the siw private drain routines makes sense
>> to me.
>> 
>> Otherwise, I think a cleaner solution is to just allow
>> processing kernel ULPs send/recv WR's while the QP is
>> already in ERROR state. In that case, we immediately
>> complete with FLUSH error. We would avoid the extra
>> state flag, and the extra check for that flag on the
>> fast path.
>> 
>> I can send such patch tomorrow if you like.
>Sure Bernard, it's good if we can avoiding extra check in fast path.
>The only condition for using ib_drain_cq(with special CQE) is: the
>special CQE should be the last CQE to be processed in the completion
>queue.
>
>Also, we can't miss the special CQE due to down_read_trylock()
>failure
>in post_send() and post_recv() routines.
>Currenlty, special CQEs are being sent only once.
>

Well, at NVMeF target side. I have seen even two consecutive
drain_sq() calls on the same QP, which does not hurt us.

I resend as v2 and have you as 'Signed-off' as well. I hope
it is the right way to signal I partially re-wrote the patch.


Many thanks,
Bernard.


>Thanks,
>Krishna.
>> 
>> Many thanks,
>> Bernard.
>> 
>> >> 
>> >> Many thanks!
>> >> Bernard.
>> >> 
>> >> 
>> >> 
>> >> 
>> >> 
>> >> >
>> >> >[  +0.001831] general protection fault: 0000 [#1] SMP PTI
>> >> >[  +0.000002] Call Trace:
>> >> >[  +0.000026]  ? __ib_process_cq+0x7a/0xc0 [ib_core]
>> >> >[  +0.000008]  ? ib_poll_handler+0x2c/0x80 [ib_core]
>> >> >[  +0.000005]  ? irq_poll_softirq+0xae/0x110
>> >> >[  +0.000005]  ? __do_softirq+0xda/0x2a8
>> >> >[  +0.000004]  ? run_ksoftirqd+0x26/0x40
>> >> >[  +0.000005]  ? smpboot_thread_fn+0x10e/0x160
>> >> >[  +0.000004]  ? kthread+0xf8/0x130
>> >> >[  +0.000003]  ? sort_range+0x20/0x20
>> >> >[  +0.000003]  ? kthread_bind+0x10/0x10
>> >> >[  +0.000003]  ? ret_from_fork+0x35/0x40
>> >> >
>> >> >Hence, changing the SQ & RQ drain logic to support current
>IB/core
>> >> >drain semantics, though this drain method does not naturally
>> >aligns
>> >> >to iWARP spec(where post_send/recv calls are not allowed in
>> >> >QP ERROR state). More on this was described in below commit:
>> >> >commit 4fe7c2962e11 ("iw_cxgb4: refactor sq/rq drain logic")
>> >> >
>> >> >Signed-off-by: Krishnamraju Eraparaju <krishna2@xxxxxxxxxxx>
>> >> >---
>> >> > drivers/infiniband/sw/siw/siw.h       |  3 +-
>> >> > drivers/infiniband/sw/siw/siw_cm.c    |  4 +-
>> >> > drivers/infiniband/sw/siw/siw_main.c  | 20 ---------
>> >> > drivers/infiniband/sw/siw/siw_verbs.c | 60
>> >> >+++++++++++++++++++++++++++
>> >> > 4 files changed, 64 insertions(+), 23 deletions(-)
>> >> >
>> >> >diff --git a/drivers/infiniband/sw/siw/siw.h
>> >> >b/drivers/infiniband/sw/siw/siw.h
>> >> >index dba4535494ab..ad4f078e4587 100644
>> >> >--- a/drivers/infiniband/sw/siw/siw.h
>> >> >+++ b/drivers/infiniband/sw/siw/siw.h
>> >> >@@ -240,7 +240,8 @@ enum siw_qp_flags {
>> >> > 	SIW_RDMA_READ_ENABLED = (1 << 2),
>> >> > 	SIW_SIGNAL_ALL_WR = (1 << 3),
>> >> > 	SIW_MPA_CRC = (1 << 4),
>> >> >-	SIW_QP_IN_DESTROY = (1 << 5)
>> >> >+	SIW_QP_IN_DESTROY = (1 << 5),
>> >> >+	SIW_QP_DRAINED_FINAL = (1 << 6)
>> >> > };
>> >> > 
>> >> > enum siw_qp_attr_mask {
>> >> >diff --git a/drivers/infiniband/sw/siw/siw_cm.c
>> >> >b/drivers/infiniband/sw/siw/siw_cm.c
>> >> >index 8c1931a57f4a..fb830622d32e 100644
>> >> >--- a/drivers/infiniband/sw/siw/siw_cm.c
>> >> >+++ b/drivers/infiniband/sw/siw/siw_cm.c
>> >> >@@ -857,7 +857,7 @@ static int siw_proc_mpareply(struct siw_cep
>> >*cep)
>> >> > 	memset(&qp_attrs, 0, sizeof(qp_attrs));
>> >> > 
>> >> > 	if (rep->params.bits & MPA_RR_FLAG_CRC)
>> >> >-		qp_attrs.flags = SIW_MPA_CRC;
>> >> >+		qp_attrs.flags |= SIW_MPA_CRC;
>> >> > 
>> >> > 	qp_attrs.irq_size = cep->ird;
>> >> > 	qp_attrs.orq_size = cep->ord;
>> >> >@@ -1675,7 +1675,7 @@ int siw_accept(struct iw_cm_id *id,
>struct
>> >> >iw_cm_conn_param *params)
>> >> > 	qp_attrs.irq_size = cep->ird;
>> >> > 	qp_attrs.sk = cep->sock;
>> >> > 	if (cep->mpa.hdr.params.bits & MPA_RR_FLAG_CRC)
>> >> >-		qp_attrs.flags = SIW_MPA_CRC;
>> >> >+		qp_attrs.flags |= SIW_MPA_CRC;
>> >> > 	qp_attrs.state = SIW_QP_STATE_RTS;
>> >> > 
>> >> > 	siw_dbg_cep(cep, "[QP%u]: moving to rts\n", qp_id(qp));
>> >> >diff --git a/drivers/infiniband/sw/siw/siw_main.c
>> >> >b/drivers/infiniband/sw/siw/siw_main.c
>> >> >index 05a92f997f60..fb01407a310f 100644
>> >> >--- a/drivers/infiniband/sw/siw/siw_main.c
>> >> >+++ b/drivers/infiniband/sw/siw/siw_main.c
>> >> >@@ -248,24 +248,6 @@ static struct ib_qp
>*siw_get_base_qp(struct
>> >> >ib_device *base_dev, int id)
>> >> > 	return NULL;
>> >> > }
>> >> > 
>> >> >-static void siw_verbs_sq_flush(struct ib_qp *base_qp)
>> >> >-{
>> >> >-	struct siw_qp *qp = to_siw_qp(base_qp);
>> >> >-
>> >> >-	down_write(&qp->state_lock);
>> >> >-	siw_sq_flush(qp);
>> >> >-	up_write(&qp->state_lock);
>> >> >-}
>> >> >-
>> >> >-static void siw_verbs_rq_flush(struct ib_qp *base_qp)
>> >> >-{
>> >> >-	struct siw_qp *qp = to_siw_qp(base_qp);
>> >> >-
>> >> >-	down_write(&qp->state_lock);
>> >> >-	siw_rq_flush(qp);
>> >> >-	up_write(&qp->state_lock);
>> >> >-}
>> >> >-
>> >> > static const struct ib_device_ops siw_device_ops = {
>> >> > 	.owner = THIS_MODULE,
>> >> > 	.uverbs_abi_ver = SIW_ABI_VERSION,
>> >> >@@ -284,8 +266,6 @@ static const struct ib_device_ops
>> >siw_device_ops
>> >> >= {
>> >> > 	.destroy_cq = siw_destroy_cq,
>> >> > 	.destroy_qp = siw_destroy_qp,
>> >> > 	.destroy_srq = siw_destroy_srq,
>> >> >-	.drain_rq = siw_verbs_rq_flush,
>> >> >-	.drain_sq = siw_verbs_sq_flush,
>> >> > 	.get_dma_mr = siw_get_dma_mr,
>> >> > 	.get_port_immutable = siw_get_port_immutable,
>> >> > 	.iw_accept = siw_accept,
>> >> >diff --git a/drivers/infiniband/sw/siw/siw_verbs.c
>> >> >b/drivers/infiniband/sw/siw/siw_verbs.c
>> >> >index 869e02b69a01..5dd62946a649 100644
>> >> >--- a/drivers/infiniband/sw/siw/siw_verbs.c
>> >> >+++ b/drivers/infiniband/sw/siw/siw_verbs.c
>> >> >@@ -596,6 +596,13 @@ int siw_verbs_modify_qp(struct ib_qp
>> >*base_qp,
>> >> >struct ib_qp_attr *attr,
>> >> > 
>> >> > 	rv = siw_qp_modify(qp, &new_attrs, siw_attr_mask);
>> >> > 
>> >> >+	/* QP state ERROR here ensures that all the SQ & RQ entries
>got
>> >> >drained
>> >> >+	 * completely. And henceforth, no more entries will be added
>to
>> >the
>> >> >CQ,
>> >> >+	 * exception is special drain CQEs via ib_drain_qp().
>> >> >+	 */
>> >> >+	if (qp->attrs.state == SIW_QP_STATE_ERROR)
>> >> >+		qp->attrs.flags |= SIW_QP_DRAINED_FINAL;
>> >> >+
>> >> > 	up_write(&qp->state_lock);
>> >> > out:
>> >> > 	return rv;
>> >> >@@ -687,6 +694,44 @@ static int siw_copy_inline_sgl(const
>struct
>> >> >ib_send_wr *core_wr,
>> >> > 	return bytes;
>> >> > }
>> >> > 
>> >> >+/* SQ final completion routine to support ib_drain_sp(). */
>> >> >+int siw_sq_final_comp(struct siw_qp *qp, const struct
>ib_send_wr
>> >> >*wr,
>> >> >+		      const struct ib_send_wr **bad_wr)
>> >> >+{
>> >> >+	struct siw_sqe sqe = {};
>> >> >+	int rv = 0;
>> >> >+
>> >> >+	while (wr) {
>> >> >+		sqe.id = wr->wr_id;
>> >> >+		sqe.opcode = wr->opcode;
>> >> >+		rv = siw_sqe_complete(qp, &sqe, 0, SIW_WC_WR_FLUSH_ERR);
>> >> >+		if (rv) {
>> >> >+			*bad_wr = wr;
>> >> >+			break;
>> >> >+		}
>> >> >+		wr = wr->next;
>> >> >+	}
>> >> >+	return rv;
>> >> >+}
>> >> >+
>> >> >+/* RQ final completion routine to support ib_drain_rp(). */
>> >> >+int siw_rq_final_comp(struct siw_qp *qp, const struct
>ib_recv_wr
>> >> >*wr,
>> >> >+		      const struct ib_recv_wr **bad_wr)
>> >> >+{
>> >> >+	struct siw_rqe rqe = {};
>> >> >+	int rv = 0;
>> >> >+
>> >> >+	while (wr) {
>> >> >+		rqe.id = wr->wr_id;
>> >> >+		rv = siw_rqe_complete(qp, &rqe, 0, 0, SIW_WC_WR_FLUSH_ERR);
>> >> >+		if (rv) {
>> >> >+			*bad_wr = wr;
>> >> >+			break;
>> >> >+		}
>> >> >+		wr = wr->next;
>> >> >+	}
>> >> >+	return rv;
>> >> >+}
>> >> > /*
>> >> >  * siw_post_send()
>> >> >  *
>> >> >@@ -705,6 +750,15 @@ int siw_post_send(struct ib_qp *base_qp,
>> >const
>> >> >struct ib_send_wr *wr,
>> >> > 	unsigned long flags;
>> >> > 	int rv = 0;
>> >> > 
>> >> >+	/* Currently there is no way to distinguish between special
>> >drain
>> >> >+	 * WRs and normal WRs(?), so we do FLUSH_ERR for all the WRs
>> >> >that've
>> >> >+	 * arrived in the ERROR/SIW_QP_DRAINED_FINAL state, assuming
>we
>> >get
>> >> >+	 * only special drain WRs in this state via ib_drain_sq().
>> >> >+	 */
>> >> >+	if (qp->attrs.flags & SIW_QP_DRAINED_FINAL) {
>> >> >+		rv = siw_sq_final_comp(qp, wr, bad_wr);
>> >> >+		return rv;
>> >> >+	}
>> >> > 	/*
>> >> > 	 * Try to acquire QP state lock. Must be non-blocking
>> >> > 	 * to accommodate kernel clients needs.
>> >> >@@ -919,6 +973,12 @@ int siw_post_receive(struct ib_qp
>*base_qp,
>> >> >const struct ib_recv_wr *wr,
>> >> > 		*bad_wr = wr;
>> >> > 		return -EOPNOTSUPP; /* what else from errno.h? */
>> >> > 	}
>> >> >+
>> >> >+	if (qp->attrs.flags & SIW_QP_DRAINED_FINAL) {
>> >> >+		rv = siw_rq_final_comp(qp, wr, bad_wr);
>> >> >+		return rv;
>> >> >+	}
>> >> >+
>> >> > 	/*
>> >> > 	 * Try to acquire QP state lock. Must be non-blocking
>> >> > 	 * to accommodate kernel clients needs.
>> >> >-- 
>> >> >2.23.0.rc0
>> >> >
>> >> >
>> >> 
>> >
>> >
>> 
>
>




[Index of Archives]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Photo]     [Yosemite News]     [Yosemite Photos]     [Linux Kernel]     [Linux SCSI]     [XFree86]

  Powered by Linux