-----"Michal Kalderon" <michal.kalderon@xxxxxxxxxxx> wrote: ----- >To: <michal.kalderon@xxxxxxxxxxx>, <ariel.elior@xxxxxxxxxxx>, ><dledford@xxxxxxxxxx>, <jgg@xxxxxxxx>, <galpress@xxxxxxxxxx>, ><yishaih@xxxxxxxxxxxx>, <bmt@xxxxxxxxxxxxxx> >From: "Michal Kalderon" <michal.kalderon@xxxxxxxxxxx> >Date: 10/30/2019 10:47AM >Cc: <linux-rdma@xxxxxxxxxxxxxxx> >Subject: [EXTERNAL] [PATCH v12 rdma-next 5/8] RDMA/siw: Use the >common mmap_xa helpers > >Remove the functions related to managing the mmap_xa database. >This code is now common in ib_core. Use the common API's instead. > >Signed-off-by: Ariel Elior <ariel.elior@xxxxxxxxxxx> >Signed-off-by: Michal Kalderon <michal.kalderon@xxxxxxxxxxx> >Signed-off-by: Bernard Metzler <bmt@xxxxxxxxxxxxxx> >--- > drivers/infiniband/sw/siw/siw.h | 20 +++- > drivers/infiniband/sw/siw/siw_main.c | 1 + > drivers/infiniband/sw/siw/siw_verbs.c | 207 >++++++++++++++++++---------------- > drivers/infiniband/sw/siw/siw_verbs.h | 1 + > 4 files changed, 128 insertions(+), 101 deletions(-) > >diff --git a/drivers/infiniband/sw/siw/siw.h >b/drivers/infiniband/sw/siw/siw.h >index dba4535494ab..824b11d5b879 100644 >--- a/drivers/infiniband/sw/siw/siw.h >+++ b/drivers/infiniband/sw/siw/siw.h >@@ -220,7 +220,7 @@ struct siw_cq { > u32 cq_get; > u32 num_cqe; > bool kernel_verbs; >- u32 xa_cq_index; /* mmap information for CQE array */ >+ struct rdma_user_mmap_entry *cq_entry; /* mmap info for CQE array >*/ > u32 id; /* For debugging only */ > }; > >@@ -263,7 +263,7 @@ struct siw_srq { > u32 rq_put; > u32 rq_get; > u32 num_rqe; /* max # of wqe's allowed */ >- u32 xa_srq_index; /* mmap information for SRQ array */ >+ struct rdma_user_mmap_entry *srq_entry; /* mmap info for SRQ array >*/ > char armed; /* inform user if limit hit */ > char kernel_verbs; /* '1' if kernel client */ > }; >@@ -477,8 +477,8 @@ struct siw_qp { > u8 layer : 4, etype : 4; > u8 ecode; > } term_info; >- u32 xa_sq_index; /* mmap information for SQE array */ >- u32 xa_rq_index; /* mmap information for RQE array */ >+ struct rdma_user_mmap_entry *sq_entry; /* mmap info for SQE array >*/ >+ struct rdma_user_mmap_entry *rq_entry; /* mmap info for RQE array >*/ > struct rcu_head rcu; > }; > >@@ -503,6 +503,12 @@ struct iwarp_msg_info { > int (*rx_data)(struct siw_qp *qp); > }; > >+struct siw_user_mmap_entry { >+ struct rdma_user_mmap_entry rdma_entry; >+ void *address; >+ size_t length; >+}; >+ > /* Global siw parameters. Currently set in siw_main.c */ > extern const bool zcopy_tx; > extern const bool try_gso; >@@ -607,6 +613,12 @@ static inline struct siw_mr *to_siw_mr(struct >ib_mr *base_mr) > return container_of(base_mr, struct siw_mr, base_mr); > } > >+static inline struct siw_user_mmap_entry * >+to_siw_mmap_entry(struct rdma_user_mmap_entry *rdma_mmap) >+{ >+ return container_of(rdma_mmap, struct siw_user_mmap_entry, >rdma_entry); >+} >+ > static inline struct siw_qp *siw_qp_id2obj(struct siw_device *sdev, >int id) > { > struct siw_qp *qp; >diff --git a/drivers/infiniband/sw/siw/siw_main.c >b/drivers/infiniband/sw/siw/siw_main.c >index d1a1b7aa7d83..4b44d5f075fd 100644 >--- a/drivers/infiniband/sw/siw/siw_main.c >+++ b/drivers/infiniband/sw/siw/siw_main.c >@@ -299,6 +299,7 @@ static const struct ib_device_ops siw_device_ops >= { > .iw_rem_ref = siw_qp_put_ref, > .map_mr_sg = siw_map_mr_sg, > .mmap = siw_mmap, >+ .mmap_free = siw_mmap_free, > .modify_qp = siw_verbs_modify_qp, > .modify_srq = siw_modify_srq, > .poll_cq = siw_poll_cq, >diff --git a/drivers/infiniband/sw/siw/siw_verbs.c >b/drivers/infiniband/sw/siw/siw_verbs.c >index 869e02b69a01..71c7a61cbfe3 100644 >--- a/drivers/infiniband/sw/siw/siw_verbs.c >+++ b/drivers/infiniband/sw/siw/siw_verbs.c >@@ -34,44 +34,20 @@ static char ib_qp_state_to_string[IB_QPS_ERR + >1][sizeof("RESET")] = { > [IB_QPS_ERR] = "ERR" > }; > >-static u32 siw_create_uobj(struct siw_ucontext *uctx, void *vaddr, >u32 size) >+void siw_mmap_free(struct rdma_user_mmap_entry *rdma_entry) > { >- struct siw_uobj *uobj; >- struct xa_limit limit = XA_LIMIT(0, SIW_UOBJ_MAX_KEY); >- u32 key; >+ struct siw_user_mmap_entry *entry = to_siw_mmap_entry(rdma_entry); > >- uobj = kzalloc(sizeof(*uobj), GFP_KERNEL); >- if (!uobj) >- return SIW_INVAL_UOBJ_KEY; >- >- if (xa_alloc_cyclic(&uctx->xa, &key, uobj, limit, >&uctx->uobj_nextkey, >- GFP_KERNEL) < 0) { >- kfree(uobj); >- return SIW_INVAL_UOBJ_KEY; >- } >- uobj->size = PAGE_ALIGN(size); >- uobj->addr = vaddr; >- >- return key; >-} >- >-static struct siw_uobj *siw_get_uobj(struct siw_ucontext *uctx, >- unsigned long off, u32 size) >-{ >- struct siw_uobj *uobj = xa_load(&uctx->xa, off); >- >- if (uobj && uobj->size == size) >- return uobj; >- >- return NULL; >+ kfree(entry); > } > > int siw_mmap(struct ib_ucontext *ctx, struct vm_area_struct *vma) > { > struct siw_ucontext *uctx = to_siw_ctx(ctx); >- struct siw_uobj *uobj; >- unsigned long off = vma->vm_pgoff; >- int size = vma->vm_end - vma->vm_start; >+ unsigned long off = vma->vm_pgoff << PAGE_SHIFT; >+ size_t size = vma->vm_end - vma->vm_start; >+ struct rdma_user_mmap_entry *rdma_entry; >+ struct siw_user_mmap_entry *entry; > int rv = -EINVAL; > > /* >@@ -79,18 +55,32 @@ int siw_mmap(struct ib_ucontext *ctx, struct >vm_area_struct *vma) > */ > if (vma->vm_start & (PAGE_SIZE - 1)) { > pr_warn("siw: mmap not page aligned\n"); >- goto out; >+ return -EINVAL; > } >- uobj = siw_get_uobj(uctx, off, size); >- if (!uobj) { >- siw_dbg(&uctx->sdev->base_dev, "mmap lookup failed: %lu, %u\n", >+ rdma_entry = rdma_user_mmap_entry_get(&uctx->base_ucontext, off, >+ vma); >+ if (!rdma_entry) { >+ siw_dbg(&uctx->sdev->base_dev, "mmap lookup failed: %lu, %#zx\n", > off, size); >+ return -EINVAL; >+ } >+ entry = to_siw_mmap_entry(rdma_entry); >+ if (PAGE_ALIGN(entry->length) != size) { >+ siw_dbg(&uctx->sdev->base_dev, >+ "key[%#lx] does not have valid length[%#zx] expected[%#zx]\n", >+ off, size, entry->length); >+ rv = -EINVAL; >+ goto out; >+ } >+ >+ rv = remap_vmalloc_range(vma, entry->address, 0); >+ if (rv) { >+ pr_warn("remap_vmalloc_range failed: %lu, %zu\n", off, size); > goto out; > } >- rv = remap_vmalloc_range(vma, uobj->addr, 0); >- if (rv) >- pr_warn("remap_vmalloc_range failed: %lu, %u\n", off, size); > out: >+ rdma_user_mmap_entry_put(&uctx->base_ucontext, rdma_entry); >+ > return rv; > } > >@@ -105,7 +95,7 @@ int siw_alloc_ucontext(struct ib_ucontext >*base_ctx, struct ib_udata *udata) > rv = -ENOMEM; > goto err_out; > } >- xa_init_flags(&ctx->xa, XA_FLAGS_ALLOC); >+ > ctx->uobj_nextkey = 0; > ctx->sdev = sdev; > >@@ -135,19 +125,7 @@ int siw_alloc_ucontext(struct ib_ucontext >*base_ctx, struct ib_udata *udata) > void siw_dealloc_ucontext(struct ib_ucontext *base_ctx) > { > struct siw_ucontext *uctx = to_siw_ctx(base_ctx); >- void *entry; >- unsigned long index; > >- /* >- * Make sure all user mmap objects are gone. Since QP, CQ >- * and SRQ destroy routines destroy related objects, nothing >- * should be found here. >- */ >- xa_for_each(&uctx->xa, index, entry) { >- kfree(xa_erase(&uctx->xa, index)); >- pr_warn("siw: dropping orphaned uobj at %lu\n", index); >- } >- xa_destroy(&uctx->xa); > atomic_dec(&uctx->sdev->num_ctx); > } > >@@ -293,6 +271,34 @@ void siw_qp_put_ref(struct ib_qp *base_qp) > siw_qp_put(to_siw_qp(base_qp)); > } > >+static struct rdma_user_mmap_entry * >+siw_mmap_entry_insert(struct siw_ucontext *uctx, >+ void *address, size_t length, >+ u64 *key) >+{ >+ struct siw_user_mmap_entry *entry = kzalloc(sizeof(*entry), >GFP_KERNEL); >+ int rv; >+ >+ *key = SIW_INVAL_UOBJ_KEY; >+ if (!entry) >+ return NULL; >+ >+ entry->address = address; >+ entry->length = length; >+ >+ rv = rdma_user_mmap_entry_insert(&uctx->base_ucontext, >+ &entry->rdma_entry, >+ length); >+ if (rv) { >+ kfree(entry); >+ return NULL; >+ } >+ >+ *key = rdma_user_mmap_get_key(&entry->rdma_entry); >+ >+ return &entry->rdma_entry; >+} >+ > /* > * siw_create_qp() > * >@@ -317,6 +323,7 @@ struct ib_qp *siw_create_qp(struct ib_pd *pd, > struct siw_cq *scq = NULL, *rcq = NULL; > unsigned long flags; > int num_sqe, num_rqe, rv = 0; >+ size_t length; > > siw_dbg(base_dev, "create new QP\n"); > >@@ -380,8 +387,6 @@ struct ib_qp *siw_create_qp(struct ib_pd *pd, > spin_lock_init(&qp->orq_lock); > > qp->kernel_verbs = !udata; >- qp->xa_sq_index = SIW_INVAL_UOBJ_KEY; >- qp->xa_rq_index = SIW_INVAL_UOBJ_KEY; > > rv = siw_qp_add(sdev, qp); > if (rv) >@@ -458,22 +463,27 @@ struct ib_qp *siw_create_qp(struct ib_pd *pd, > uresp.qp_id = qp_id(qp); > > if (qp->sendq) { >- qp->xa_sq_index = >- siw_create_uobj(uctx, qp->sendq, >- num_sqe * sizeof(struct siw_sqe)); >+ length = num_sqe * sizeof(struct siw_sqe); >+ qp->sq_entry = >+ siw_mmap_entry_insert(uctx, qp->sendq, >+ length, &uresp.sq_key); >+ if (!qp->sq_entry) { >+ rv = -ENOMEM; >+ goto err_out_xa; >+ } > } >+ > if (qp->recvq) { >- qp->xa_rq_index = >- siw_create_uobj(uctx, qp->recvq, >- num_rqe * sizeof(struct siw_rqe)); >- } >- if (qp->xa_sq_index == SIW_INVAL_UOBJ_KEY || >- qp->xa_rq_index == SIW_INVAL_UOBJ_KEY) { >- rv = -ENOMEM; >- goto err_out_xa; >+ length = num_rqe * sizeof(struct siw_rqe); >+ qp->rq_entry = >+ siw_mmap_entry_insert(uctx, qp->recvq, >+ length, &uresp.rq_key); >+ if (!qp->rq_entry) { >+ uresp.sq_key = SIW_INVAL_UOBJ_KEY; >+ rv = -ENOMEM; >+ goto err_out_xa; >+ } > } >- uresp.sq_key = qp->xa_sq_index << PAGE_SHIFT; >- uresp.rq_key = qp->xa_rq_index << PAGE_SHIFT; > > if (udata->outlen < sizeof(uresp)) { > rv = -EINVAL; >@@ -501,11 +511,12 @@ struct ib_qp *siw_create_qp(struct ib_pd *pd, > kfree(siw_base_qp); > > if (qp) { >- if (qp->xa_sq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&uctx->xa, qp->xa_sq_index)); >- if (qp->xa_rq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&uctx->xa, qp->xa_rq_index)); >- >+ if (uctx) { >+ rdma_user_mmap_entry_remove(&uctx->base_ucontext, >+ qp->sq_entry); >+ rdma_user_mmap_entry_remove(&uctx->base_ucontext, >+ qp->rq_entry); >+ } > vfree(qp->sendq); > vfree(qp->recvq); > kfree(qp); >@@ -619,10 +630,10 @@ int siw_destroy_qp(struct ib_qp *base_qp, >struct ib_udata *udata) > qp->attrs.flags |= SIW_QP_IN_DESTROY; > qp->rx_stream.rx_suspend = 1; > >- if (uctx && qp->xa_sq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&uctx->xa, qp->xa_sq_index)); >- if (uctx && qp->xa_rq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&uctx->xa, qp->xa_rq_index)); >+ if (uctx) { >+ rdma_user_mmap_entry_remove(&uctx->base_ucontext, qp->sq_entry); >+ rdma_user_mmap_entry_remove(&uctx->base_ucontext, qp->rq_entry); >+ } > > down_write(&qp->state_lock); > >@@ -993,8 +1004,8 @@ void siw_destroy_cq(struct ib_cq *base_cq, >struct ib_udata *udata) > > siw_cq_flush(cq); > >- if (ctx && cq->xa_cq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&ctx->xa, cq->xa_cq_index)); >+ if (ctx) >+ rdma_user_mmap_entry_remove(&ctx->base_ucontext, cq->cq_entry); > > atomic_dec(&sdev->num_cq); > >@@ -1031,7 +1042,6 @@ int siw_create_cq(struct ib_cq *base_cq, const >struct ib_cq_init_attr *attr, > size = roundup_pow_of_two(size); > cq->base_cq.cqe = size; > cq->num_cqe = size; >- cq->xa_cq_index = SIW_INVAL_UOBJ_KEY; > > if (!udata) { > cq->kernel_verbs = 1; >@@ -1057,16 +1067,17 @@ int siw_create_cq(struct ib_cq *base_cq, >const struct ib_cq_init_attr *attr, > struct siw_ucontext *ctx = > rdma_udata_to_drv_context(udata, struct siw_ucontext, > base_ucontext); >+ size_t length = size * sizeof(struct siw_cqe) + >+ sizeof(struct siw_cq_ctrl); > >- cq->xa_cq_index = >- siw_create_uobj(ctx, cq->queue, >- size * sizeof(struct siw_cqe) + >- sizeof(struct siw_cq_ctrl)); >- if (cq->xa_cq_index == SIW_INVAL_UOBJ_KEY) { >+ cq->cq_entry = >+ siw_mmap_entry_insert(ctx, cq->queue, >+ length, &uresp.cq_key); >+ if (!cq->cq_entry) { > rv = -ENOMEM; > goto err_out; > } >- uresp.cq_key = cq->xa_cq_index << PAGE_SHIFT; >+ > uresp.cq_id = cq->id; > uresp.num_cqe = size; > >@@ -1087,8 +1098,9 @@ int siw_create_cq(struct ib_cq *base_cq, const >struct ib_cq_init_attr *attr, > struct siw_ucontext *ctx = > rdma_udata_to_drv_context(udata, struct siw_ucontext, > base_ucontext); >- if (cq->xa_cq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&ctx->xa, cq->xa_cq_index)); >+ if (ctx) >+ rdma_user_mmap_entry_remove(&ctx->base_ucontext, >+ cq->cq_entry); > vfree(cq->queue); > } > atomic_dec(&sdev->num_cq); >@@ -1492,7 +1504,6 @@ int siw_create_srq(struct ib_srq *base_srq, > } > srq->max_sge = attrs->max_sge; > srq->num_rqe = roundup_pow_of_two(attrs->max_wr); >- srq->xa_srq_index = SIW_INVAL_UOBJ_KEY; > srq->limit = attrs->srq_limit; > if (srq->limit) > srq->armed = 1; >@@ -1511,15 +1522,16 @@ int siw_create_srq(struct ib_srq *base_srq, > } > if (udata) { > struct siw_uresp_create_srq uresp = {}; >+ size_t length = srq->num_rqe * sizeof(struct siw_rqe); > >- srq->xa_srq_index = siw_create_uobj( >- ctx, srq->recvq, srq->num_rqe * sizeof(struct siw_rqe)); >- >- if (srq->xa_srq_index == SIW_INVAL_UOBJ_KEY) { >+ srq->srq_entry = >+ siw_mmap_entry_insert(ctx, srq->recvq, >+ length, &uresp.srq_key); >+ if (!srq->srq_entry) { > rv = -ENOMEM; > goto err_out; > } >- uresp.srq_key = srq->xa_srq_index; >+ > uresp.num_rqe = srq->num_rqe; > > if (udata->outlen < sizeof(uresp)) { >@@ -1538,8 +1550,9 @@ int siw_create_srq(struct ib_srq *base_srq, > > err_out: > if (srq->recvq) { >- if (ctx && srq->xa_srq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&ctx->xa, srq->xa_srq_index)); >+ if (ctx) >+ rdma_user_mmap_entry_remove(&ctx->base_ucontext, >+ srq->srq_entry); > vfree(srq->recvq); > } > atomic_dec(&sdev->num_srq); >@@ -1625,9 +1638,9 @@ void siw_destroy_srq(struct ib_srq *base_srq, >struct ib_udata *udata) > rdma_udata_to_drv_context(udata, struct siw_ucontext, > base_ucontext); > >- if (ctx && srq->xa_srq_index != SIW_INVAL_UOBJ_KEY) >- kfree(xa_erase(&ctx->xa, srq->xa_srq_index)); >- >+ if (ctx) >+ rdma_user_mmap_entry_remove(&ctx->base_ucontext, >+ srq->srq_entry); > vfree(srq->recvq); > atomic_dec(&sdev->num_srq); > } >diff --git a/drivers/infiniband/sw/siw/siw_verbs.h >b/drivers/infiniband/sw/siw/siw_verbs.h >index 1910869281cb..1a731989fad6 100644 >--- a/drivers/infiniband/sw/siw/siw_verbs.h >+++ b/drivers/infiniband/sw/siw/siw_verbs.h >@@ -83,6 +83,7 @@ void siw_destroy_srq(struct ib_srq *base_srq, >struct ib_udata *udata); > int siw_post_srq_recv(struct ib_srq *base_srq, const struct >ib_recv_wr *wr, > const struct ib_recv_wr **bad_wr); > int siw_mmap(struct ib_ucontext *ctx, struct vm_area_struct *vma); >+void siw_mmap_free(struct rdma_user_mmap_entry *rdma_entry); > void siw_qp_event(struct siw_qp *qp, enum ib_event_type type); > void siw_cq_event(struct siw_cq *cq, enum ib_event_type type); > void siw_srq_event(struct siw_srq *srq, enum ib_event_type type); >-- >2.14.5 > > Hi Michal, For the siw part, this patch works fine and looks good. Reviewed-by: Bernard Metzler <bmt@xxxxxxxxxxxxxx> Tested-by: Bernard Metzler <bmt@xxxxxxxxxxxxxx>