Re: Question about nfs in infiniband environment

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Aug 28, 2018 at 11:41 AM Chuck Lever <chuck.lever@xxxxxxxxxx> wrote:
>
>
>
> > On Aug 28, 2018, at 11:31 AM, Volker Lieder <v.lieder@xxxxxxxxxx> wrote:
> >
> > Hi Chuck,
> >
> >> Am 28.08.2018 um 17:26 schrieb Chuck Lever <chucklever@xxxxxxxxx>:
> >>
> >> Hi Volker-
> >>
> >>
> >>> On Aug 28, 2018, at 8:37 AM, Volker Lieder <v.lieder@xxxxxxxxxx> wrote:
> >>>
> >>> Hi,
> >>>
> >>> a short update from our site.
> >>>
> >>> We resized CPU and RAM on the nfs server and the performance is good right now and the error messages are gone.
> >>>
> >>> Is there a guide what hardware requirements a fast nfs server has?
> >>>
> >>> Or an information, how many nfs prozesses are needed for x nfs clients?
> >>
> >> The nfsd thread count depends on number of clients _and_ their workload.
> >> There isn't a hard and fast rule.
> >>
> >> The default thread count is probably too low for your workload. You can
> >> edit /etc/sysconfig/nfs and find "RPCNFSDCOUNT". Increase it to, say,
> >> 64, and restart your NFS server.
> >
> > I tried this, but then the load on the "small" server was to high to serve further requests, so that was the idea to grow this up.
>
> That rather suggests the disks are slow. A deeper performance
> analysis might help.
>
>
> >> With InfiniBand you also have the option of using NFS/RDMA. Mount with
> >> "proto=rdma,port=20049" to try it.
> >
> > Yes, thats true, but in the mellanox driver set they disabled nfsordma in Version 3.4.
>
> Not quite sure what you mean by "mellanox driver". Do you
> mean MOFED? My impression of the stock CentOS 7.5 code is
> that it is close to upstream, and you shouldn't need to
> replace it except in some very special circumstances (high
> end database, eg).
>
>
> > It should work with centos driver, but we didnt tested it right now in newer setups.
> >
> > One more question, since other problems seem to be solved:
> >
> > What about this message?
> >
> > [Tue Aug 28 15:10:44 2018] NFSD: client 172.16.YY.XXX testing state ID with incorrect client ID
>
> Looks like an NFS bug. Someone else on the list should be able
> to comment.

I ran into this problem while testing RHEL7.5 NFSoRDMA (over
SoftRoCE). Here's a bugzilla
https://bugzilla.redhat.com/show_bug.cgi?id=1518006

I was having a hard time reproducing it consistently to debug it.
Because it was really a non-error error (and it wasn't upstream), it
went on a back burner.

>
>
> >>> Best regards,
> >>> Volker
> >>>
> >>>> Am 28.08.2018 um 09:45 schrieb Volker Lieder <v.lieder@xxxxxxxxxx>:
> >>>>
> >>>> Hi list,
> >>>>
> >>>> we have a setup with round about 15 centos 7.5 server.
> >>>>
> >>>> All are connected via infiniband 56Gbit and installed with new mellanox driver.
> >>>> One server (4 Core, 8 threads, 16GB) is nfs server for a disk shelf with round about 500TB data.
> >>>>
> >>>> The server exports 4-6 mounts to each client.
> >>>>
> >>>> Since we added 3 further nodes to the setup, we recieve following messages:
> >>>>
> >>>> On nfs-server:
> >>>> [Tue Aug 28 07:29:33 2018] rpc-srv/tcp: nfsd: sent only 224000 when sending 1048684 bytes - shutting down socket
> >>>> [Tue Aug 28 07:30:13 2018] rpc-srv/tcp: nfsd: sent only 209004 when sending 1048684 bytes - shutting down socket
> >>>> [Tue Aug 28 07:30:14 2018] rpc-srv/tcp: nfsd: sent only 204908 when sending 630392 bytes - shutting down socket
> >>>> [Tue Aug 28 07:32:31 2018] rpc-srv/tcp: nfsd: got error -11 when sending 524396 bytes - shutting down socket
> >>>> [Tue Aug 28 07:32:33 2018] rpc-srv/tcp: nfsd: got error -11 when sending 308 bytes - shutting down socket
> >>>> [Tue Aug 28 07:32:35 2018] rpc-srv/tcp: nfsd: got error -11 when sending 172 bytes - shutting down socket
> >>>> [Tue Aug 28 07:32:53 2018] rpc-srv/tcp: nfsd: got error -11 when sending 164 bytes - shutting down socket
> >>>> [Tue Aug 28 07:38:52 2018] rpc-srv/tcp: nfsd: sent only 749452 when sending 1048684 bytes - shutting down socket
> >>>> [Tue Aug 28 07:39:29 2018] rpc-srv/tcp: nfsd: got error -11 when sending 244 bytes - shutting down socket
> >>>> [Tue Aug 28 07:39:29 2018] rpc-srv/tcp: nfsd: got error -11 when sending 1048684 bytes - shutting down socket
> >>>>
> >>>> on nfs-clients:
> >>>> [229903.273435] nfs: server 172.16.55.221 not responding, still trying
> >>>> [229903.523455] nfs: server 172.16.55.221 OK
> >>>> [229939.080276] nfs: server 172.16.55.221 OK
> >>>> [236527.473064] perf: interrupt took too long (6226 > 6217), lowering kernel.perf_event_max_sample_rate to 32000
> >>>> [248874.777322] RPC: Could not send backchannel reply error: -105
> >>>> [249484.823793] RPC: Could not send backchannel reply error: -105
> >>>> [250382.497448] RPC: Could not send backchannel reply error: -105
> >>>> [250671.054112] RPC: Could not send backchannel reply error: -105
> >>>> [251284.622707] RPC: Could not send backchannel reply error: -105
> >>>>
> >>>> Also file requests or "df -h" ended sometimes in a stale nfs status whcih will be good after a minute.
> >>>>
> >>>> I googled all messages and tried different things without success.
> >>>> We are now going on to upgrade cpu power on nfs server.
> >>>>
> >>>> Do you also have any hints or points i can look for?
> >>>>
> >>>> Best regards,
> >>>> Volker
> >>>
> >>
> >> --
> >> Chuck Lever
> >> chucklever@xxxxxxxxx
>
> --
> Chuck Lever
>
>
>



[Index of Archives]     [Linux Filesystem Development]     [Linux USB Development]     [Linux Media Development]     [Video for Linux]     [Linux NILFS]     [Linux Audio Users]     [Yosemite Info]     [Linux SCSI]

  Powered by Linux