Re: [PATCH REPOST2] svcrdma: Limit ORD based on client's advertised IRD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, 2008-05-20 at 17:24 -0400, J. Bruce Fields wrote:
> On Mon, May 19, 2008 at 07:11:35PM -0500, Tom Tucker wrote:
> > When adapters have differing IRD limits, the RDMA transport will fail to
> > connect properly.
>  
> How will this look to the user?  Are they going to get some sort of
> mysterious failure when they try to mount, and then have to go echo
> lower values into /proc/sys/sunrpc/svc_rdma/max_outbound_read_requests
> until it works?

BTW, this is a hypothetical for iWARP. The last I checked, the current
spate of adapters seemed to work fine (Chelsio/Neteffect) because the
NFSRDMA default is supported by both. I'll check again and get back to
you.

Tom

> 
> > The RDMA transport should use the client's advertised
> > inbound read limit when computing its outbound read limit. For iWARP
> > transports, there is currently no standard for exchanging IRD/ORD
> > during connection establishment so the 'responder_resources' field in the
> > connect event is the local device's limit. The RDMA transport can be
> > configured to use a smaller ORD by writing the desired number to the
> > /proc/sys/sunrpc/svc_rdma/max_outbound_read_requests file.
> 
> Do we have any documentation on using these various sysctls?
> 
> --b.
> 
> >     
> > Signed-off-by: Tom Tucker <tom@xxxxxxxxxxxxxxxxxxxxx>
> > ---
> > 
> > This one includes the min_t change. This is nicer. Thanks Tom.
> > 
> >  net/sunrpc/xprtrdma/svc_rdma_transport.c |   16 ++++++++++++----
> >  1 files changed, 12 insertions(+), 4 deletions(-)
> > 
> > diff --git a/net/sunrpc/xprtrdma/svc_rdma_transport.c b/net/sunrpc/xprtrdma/svc_rdma_transport.c
> > index 68908b5..54b2126 100644
> > --- a/net/sunrpc/xprtrdma/svc_rdma_transport.c
> > +++ b/net/sunrpc/xprtrdma/svc_rdma_transport.c
> > @@ -580,7 +580,7 @@ int svc_rdma_post_recv(struct svcxprt_rdma *xprt)
> >   * will call the recvfrom method on the listen xprt which will accept the new
> >   * connection.
> >   */
> > -static void handle_connect_req(struct rdma_cm_id *new_cma_id)
> > +static void handle_connect_req(struct rdma_cm_id *new_cma_id, size_t client_ird)
> >  {
> >  	struct svcxprt_rdma *listen_xprt = new_cma_id->context;
> >  	struct svcxprt_rdma *newxprt;
> > @@ -597,6 +597,9 @@ static void handle_connect_req(struct rdma_cm_id *new_cma_id)
> >  	dprintk("svcrdma: Creating newxprt=%p, cm_id=%p, listenxprt=%p\n",
> >  		newxprt, newxprt->sc_cm_id, listen_xprt);
> >  
> > +	/* Save client advertised inbound read limit for use later in accept. */
> > +	newxprt->sc_ord = client_ird;
> > +
> >  	/* Set the local and remote addresses in the transport */
> >  	sa = (struct sockaddr *)&newxprt->sc_cm_id->route.addr.dst_addr;
> >  	svc_xprt_set_remote(&newxprt->sc_xprt, sa, svc_addr_len(sa));
> > @@ -633,7 +636,8 @@ static int rdma_listen_handler(struct rdma_cm_id *cma_id,
> >  	case RDMA_CM_EVENT_CONNECT_REQUEST:
> >  		dprintk("svcrdma: Connect request on cma_id=%p, xprt = %p, "
> >  			"event=%d\n", cma_id, cma_id->context, event->event);
> > -		handle_connect_req(cma_id);
> > +		handle_connect_req(cma_id,
> > +				   event->param.conn.responder_resources);
> >  		break;
> >  
> >  	case RDMA_CM_EVENT_ESTABLISHED:
> > @@ -807,8 +811,12 @@ static struct svc_xprt *svc_rdma_accept(struct svc_xprt *xprt)
> >  				   (size_t)svcrdma_max_requests);
> >  	newxprt->sc_sq_depth = RPCRDMA_SQ_DEPTH_MULT * newxprt->sc_max_requests;
> >  
> > -	newxprt->sc_ord =  min((size_t)devattr.max_qp_rd_atom,
> > -			       (size_t)svcrdma_ord);
> > +	/*
> > +	 * Limit ORD based on client limit, local device limit, and
> > +	 * configured svcrdma limit.
> > +	 */
> > +	newxprt->sc_ord = min_t(size_t, devattr.max_qp_rd_atom, newxprt->sc_ord);
> > +	newxprt->sc_ord = min_t(size_t,	svcrdma_ord, newxprt->sc_ord);
> >  
> >  	newxprt->sc_pd = ib_alloc_pd(newxprt->sc_cm_id->device);
> >  	if (IS_ERR(newxprt->sc_pd)) {
> > 
> --
> To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux Filesystem Development]     [Linux USB Development]     [Linux Media Development]     [Video for Linux]     [Linux NILFS]     [Linux Audio Users]     [Yosemite Info]     [Linux SCSI]

  Powered by Linux