Re: [PATCH v9 10/25] RDMA/rtrs: server: main functionality

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Mar 05, 2020 at 01:28:39PM +0100, Jinpu Wang wrote:
> On Thu, Mar 5, 2020 at 1:16 PM Leon Romanovsky <leon@xxxxxxxxxx> wrote:
> >
> > On Thu, Mar 05, 2020 at 01:01:08PM +0100, Danil Kipnis wrote:
> > > On Thu, 5 Mar 2020, 09:00 Leon Romanovsky, <leon@xxxxxxxxxx> wrote:
> > >
> > > > On Wed, Mar 04, 2020 at 12:03:32PM +0100, Jinpu Wang wrote:
> > > > > On Tue, Mar 3, 2020 at 5:59 PM Leon Romanovsky <leon@xxxxxxxxxx> wrote:
> > > > > >
> > > > > > On Tue, Mar 03, 2020 at 05:41:27PM +0100, Jinpu Wang wrote:
> > > > > > > On Tue, Mar 3, 2020 at 12:37 PM Leon Romanovsky <leon@xxxxxxxxxx>
> > > > wrote:
> > > > > > > >
> > > > > > > > On Fri, Feb 21, 2020 at 11:47:06AM +0100, Jack Wang wrote:
> > > > > > > > > From: Jack Wang <jinpu.wang@xxxxxxxxxxxxxxx>
> > > > > > > > >
> > > > > > > > > This is main functionality of rtrs-server module, which accepts
> > > > > > > > > set of RDMA connections (so called rtrs session),
> > > > creates/destroys
> > > > > > > > > sysfs entries associated with rtrs session and notifies upper
> > > > layer
> > > > > > > > > (user of RTRS API) about RDMA requests or link events.
> > > > > > > > >
> > > > > > > > > Signed-off-by: Danil Kipnis <danil.kipnis@xxxxxxxxxxxxxxx>
> > > > > > > > > Signed-off-by: Jack Wang <jinpu.wang@xxxxxxxxxxxxxxx>
> > > > > > > > > ---
> > > > > > > > >  drivers/infiniband/ulp/rtrs/rtrs-srv.c | 2164
> > > > ++++++++++++++++++++++++
> > > > > > > > >  1 file changed, 2164 insertions(+)
> > > > > > > > >  create mode 100644 drivers/infiniband/ulp/rtrs/rtrs-srv.c
> > > > > > > > >
> > > > > > > > > diff --git a/drivers/infiniband/ulp/rtrs/rtrs-srv.c
> > > > b/drivers/infiniband/ulp/rtrs/rtrs-srv.c
> > > > > > > > > new file mode 100644
> > > > > > > > > index 000000000000..e60ee6dd675d
> > > > > > > > > --- /dev/null
> > > > > > > > > +++ b/drivers/infiniband/ulp/rtrs/rtrs-srv.c
> > > > > > > > > @@ -0,0 +1,2164 @@
> > > > > > > > > +// SPDX-License-Identifier: GPL-2.0-or-later
> > > > > > > > > +/*
> > > > > > > > > + * RDMA Transport Layer
> > > > > > > > > + *
> > > > > > > > > + * Copyright (c) 2014 - 2018 ProfitBricks GmbH. All rights
> > > > reserved.
> > > > > > > > > + * Copyright (c) 2018 - 2019 1&1 IONOS Cloud GmbH. All rights
> > > > reserved.
> > > > > > > > > + * Copyright (c) 2019 - 2020 1&1 IONOS SE. All rights reserved.
> > > > > > > > > + */
> > > > > > > > > +
> > > > > > > > > +#undef pr_fmt
> > > > > > > > > +#define pr_fmt(fmt) KBUILD_MODNAME " L" __stringify(__LINE__)
> > > > ": " fmt
> > > > > > > > > +
> > > > > > > > > +#include <linux/module.h>
> > > > > > > > > +#include <linux/mempool.h>
> > > > > > > > > +
> > > > > > > > > +#include "rtrs-srv.h"
> > > > > > > > > +#include "rtrs-log.h"
> > > > > > > > > +
> > > > > > > > > +MODULE_DESCRIPTION("RDMA Transport Server");
> > > > > > > > > +MODULE_LICENSE("GPL");
> > > > > > > > > +
> > > > > > > > > +/* Must be power of 2, see mask from mr->page_size in
> > > > ib_sg_to_pages() */
> > > > > > > > > +#define DEFAULT_MAX_CHUNK_SIZE (128 << 10)
> > > > > > > > > +#define DEFAULT_SESS_QUEUE_DEPTH 512
> > > > > > > > > +#define MAX_HDR_SIZE PAGE_SIZE
> > > > > > > > > +#define MAX_SG_COUNT ((MAX_HDR_SIZE - sizeof(struct
> > > > rtrs_msg_rdma_read)) \
> > > > > > > > > +                   / sizeof(struct rtrs_sg_desc))
> > > > > > > > > +
> > > > > > > > > +/* We guarantee to serve 10 paths at least */
> > > > > > > > > +#define CHUNK_POOL_SZ 10
> > > > > > > > > +
> > > > > > > > > +static struct rtrs_rdma_dev_pd dev_pd;
> > > > > > > > > +static mempool_t *chunk_pool;
> > > > > > > > > +struct class *rtrs_dev_class;
> > > > > > > > > +
> > > > > > > > > +static int __read_mostly max_chunk_size =
> > > > DEFAULT_MAX_CHUNK_SIZE;
> > > > > > > > > +static int __read_mostly sess_queue_depth =
> > > > DEFAULT_SESS_QUEUE_DEPTH;
> > > > > > > > > +
> > > > > > > > > +static bool always_invalidate = true;
> > > > > > > > > +module_param(always_invalidate, bool, 0444);
> > > > > > > > > +MODULE_PARM_DESC(always_invalidate,
> > > > > > > > > +              "Invalidate memory registration for contiguous
> > > > memory regions before accessing.");
> > > > > > > > > +
> > > > > > > > > +module_param_named(max_chunk_size, max_chunk_size, int, 0444);
> > > > > > > > > +MODULE_PARM_DESC(max_chunk_size,
> > > > > > > > > +              "Max size for each IO request, when change the
> > > > unit is in byte (default: "
> > > > > > > > > +              __stringify(DEFAULT_MAX_CHUNK_SIZE) "KB)");
> > > > > > > > > +
> > > > > > > > > +module_param_named(sess_queue_depth, sess_queue_depth, int,
> > > > 0444);
> > > > > > > > > +MODULE_PARM_DESC(sess_queue_depth,
> > > > > > > > > +              "Number of buffers for pending I/O requests to
> > > > allocate per session. Maximum: "
> > > > > > > > > +              __stringify(MAX_SESS_QUEUE_DEPTH) " (default: "
> > > > > > > > > +              __stringify(DEFAULT_SESS_QUEUE_DEPTH) ")");
> > > > > > > >
> > > > > > > > We don't like module parameters in the RDMA.
> > > > > > > Hi Leon,
> > > > > > >
> > > > > > > These paramters are affecting resouce usage/performance, I think
> > > > would
> > > > > > > be good to have them as module parameters,
> > > > > > > so admin could choose based their needs.
> > > > > >
> > > > > > It is premature optimization before second user comes, also it is
> > > > > > based on the assumption that everyone uses modules, which is not true.
> > > > > The idea to have module parameters is to cover more use cases, IMHO.
> > > > >
> > > > > Even you builtin the module to the kernel, you can still change the
> > > > > module parameters
> > > > > by passing the "moduls_name.paramters" in kernel command line, eg:
> > > > > kvm.nx_huge_pages=true
> > > >
> > > > I know about that, but it doesn't make them helpful.
> > > >
> > > > Thanks
> > > >
> > > Hi Leon,
> > >
> > > Queue_depth and max_chunksize parameters control the tradeoff between
> > > throuput performance and memory consumption. We do use them to set
> > > different values for storages equipped with SSDs (fast) and on storages
> > > equipped with HDDs (slow). The last parameter always_invaldate enforces the
> > > invalidation of an rdma buffer before its hand over to the block layer. We
> > > set it to no in our datacenters, since they are closed and malicious
> > > clients are not a threat in our scenario. In general case it defaults to
> > > yes, as requested by Jason. Our admins need to have control over those
> > > control knobs somehow... We could make sysfs entries out of them or
> > > something, but would it really make sense?
> >
> > blk_queue_nonrot() inside your code?
> It's exported function, and also used by other drivers like
> md/dm/target core, right?

I have no clue.

Thanks

>
> Thanks



[Index of Archives]     [Linux RAID]     [Linux SCSI]     [Linux ATA RAID]     [IDE]     [Linux Wireless]     [Linux Kernel]     [ATH6KL]     [Linux Bluetooth]     [Linux Netdev]     [Kernel Newbies]     [Security]     [Git]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Device Mapper]

  Powered by Linux