On Tue, Aug 28, 2018 at 11:41 AM Chuck Lever <chuck.lever@xxxxxxxxxx> wrote: > > > > > On Aug 28, 2018, at 11:31 AM, Volker Lieder <v.lieder@xxxxxxxxxx> wrote: > > > > Hi Chuck, > > > >> Am 28.08.2018 um 17:26 schrieb Chuck Lever <chucklever@xxxxxxxxx>: > >> > >> Hi Volker- > >> > >> > >>> On Aug 28, 2018, at 8:37 AM, Volker Lieder <v.lieder@xxxxxxxxxx> wrote: > >>> > >>> Hi, > >>> > >>> a short update from our site. > >>> > >>> We resized CPU and RAM on the nfs server and the performance is good right now and the error messages are gone. > >>> > >>> Is there a guide what hardware requirements a fast nfs server has? > >>> > >>> Or an information, how many nfs prozesses are needed for x nfs clients? > >> > >> The nfsd thread count depends on number of clients _and_ their workload. > >> There isn't a hard and fast rule. > >> > >> The default thread count is probably too low for your workload. You can > >> edit /etc/sysconfig/nfs and find "RPCNFSDCOUNT". Increase it to, say, > >> 64, and restart your NFS server. > > > > I tried this, but then the load on the "small" server was to high to serve further requests, so that was the idea to grow this up. > > That rather suggests the disks are slow. A deeper performance > analysis might help. > > > >> With InfiniBand you also have the option of using NFS/RDMA. Mount with > >> "proto=rdma,port=20049" to try it. > > > > Yes, thats true, but in the mellanox driver set they disabled nfsordma in Version 3.4. > > Not quite sure what you mean by "mellanox driver". Do you > mean MOFED? My impression of the stock CentOS 7.5 code is > that it is close to upstream, and you shouldn't need to > replace it except in some very special circumstances (high > end database, eg). > > > > It should work with centos driver, but we didnt tested it right now in newer setups. > > > > One more question, since other problems seem to be solved: > > > > What about this message? > > > > [Tue Aug 28 15:10:44 2018] NFSD: client 172.16.YY.XXX testing state ID with incorrect client ID > > Looks like an NFS bug. Someone else on the list should be able > to comment. I ran into this problem while testing RHEL7.5 NFSoRDMA (over SoftRoCE). Here's a bugzilla https://bugzilla.redhat.com/show_bug.cgi?id=1518006 I was having a hard time reproducing it consistently to debug it. Because it was really a non-error error (and it wasn't upstream), it went on a back burner. > > > >>> Best regards, > >>> Volker > >>> > >>>> Am 28.08.2018 um 09:45 schrieb Volker Lieder <v.lieder@xxxxxxxxxx>: > >>>> > >>>> Hi list, > >>>> > >>>> we have a setup with round about 15 centos 7.5 server. > >>>> > >>>> All are connected via infiniband 56Gbit and installed with new mellanox driver. > >>>> One server (4 Core, 8 threads, 16GB) is nfs server for a disk shelf with round about 500TB data. > >>>> > >>>> The server exports 4-6 mounts to each client. > >>>> > >>>> Since we added 3 further nodes to the setup, we recieve following messages: > >>>> > >>>> On nfs-server: > >>>> [Tue Aug 28 07:29:33 2018] rpc-srv/tcp: nfsd: sent only 224000 when sending 1048684 bytes - shutting down socket > >>>> [Tue Aug 28 07:30:13 2018] rpc-srv/tcp: nfsd: sent only 209004 when sending 1048684 bytes - shutting down socket > >>>> [Tue Aug 28 07:30:14 2018] rpc-srv/tcp: nfsd: sent only 204908 when sending 630392 bytes - shutting down socket > >>>> [Tue Aug 28 07:32:31 2018] rpc-srv/tcp: nfsd: got error -11 when sending 524396 bytes - shutting down socket > >>>> [Tue Aug 28 07:32:33 2018] rpc-srv/tcp: nfsd: got error -11 when sending 308 bytes - shutting down socket > >>>> [Tue Aug 28 07:32:35 2018] rpc-srv/tcp: nfsd: got error -11 when sending 172 bytes - shutting down socket > >>>> [Tue Aug 28 07:32:53 2018] rpc-srv/tcp: nfsd: got error -11 when sending 164 bytes - shutting down socket > >>>> [Tue Aug 28 07:38:52 2018] rpc-srv/tcp: nfsd: sent only 749452 when sending 1048684 bytes - shutting down socket > >>>> [Tue Aug 28 07:39:29 2018] rpc-srv/tcp: nfsd: got error -11 when sending 244 bytes - shutting down socket > >>>> [Tue Aug 28 07:39:29 2018] rpc-srv/tcp: nfsd: got error -11 when sending 1048684 bytes - shutting down socket > >>>> > >>>> on nfs-clients: > >>>> [229903.273435] nfs: server 172.16.55.221 not responding, still trying > >>>> [229903.523455] nfs: server 172.16.55.221 OK > >>>> [229939.080276] nfs: server 172.16.55.221 OK > >>>> [236527.473064] perf: interrupt took too long (6226 > 6217), lowering kernel.perf_event_max_sample_rate to 32000 > >>>> [248874.777322] RPC: Could not send backchannel reply error: -105 > >>>> [249484.823793] RPC: Could not send backchannel reply error: -105 > >>>> [250382.497448] RPC: Could not send backchannel reply error: -105 > >>>> [250671.054112] RPC: Could not send backchannel reply error: -105 > >>>> [251284.622707] RPC: Could not send backchannel reply error: -105 > >>>> > >>>> Also file requests or "df -h" ended sometimes in a stale nfs status whcih will be good after a minute. > >>>> > >>>> I googled all messages and tried different things without success. > >>>> We are now going on to upgrade cpu power on nfs server. > >>>> > >>>> Do you also have any hints or points i can look for? > >>>> > >>>> Best regards, > >>>> Volker > >>> > >> > >> -- > >> Chuck Lever > >> chucklever@xxxxxxxxx > > -- > Chuck Lever > > >