Re: NFS4 server, performance and cache consistency in comparison with solaris.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



>
> On Jul 31, 2013, at 6:35 AM, Anton Starikov <ant.starikov@xxxxxxxxx> wrote:
>
> > Hey,
> >
> > we are in the process of migration of our storage from solaris to linux (RHEL6) and I see some strange things, which depends on server side.
> >
> > In our old solaris setup we have slower network (1GbE vs 10GbE), much slower storage than new one, but still when I export volume with default solaris options and mount on linux clients with options:
> > rw,relatime,vers=4,rsize=1048576,wsize=1048576,namlen=255,hard,proto=tcp,port=0,timeo=600,retrans=2,sec=sys,minorversion=0,local_lock=none
> >

> >
> > When I export from linux host with options:
> >
> > rw,sync,wdelay,hide,nocrossmnt,insecure,no_root_squash,no_all_squash,no_subtree_check,secure_locks,no_acl,mountpoint,anonuid=65534,anongid=65534,sec=sys,rw,no_root_squash,no_all_squash (it is options as shown in /var/lib/nfs/etab),
> >
> > and mount on linux clients with the same options as with old setup, I have great "dbench 4" performance (about 200Mb/s), but consistency is nonexistent, in the same case (one client keep writing to file, second is reading), on second client I can see state of file with delayed for 5-30 secs. Out of curiosity I tried to use "sync" in loop on a first client (where it writes) to flush cache, but it does not affect something. File isn't really large, but client updates it 2-4 times a sec.
> >
> > All my attempts to improve consistency had two possible impacts:
> >
> > 1) either still luck of consistency (like actimeo=0,lookupcache=none) and reasonable or good dbench results.
> > 2) consistency is recovered (or almost recovered) (like sync, noac), but dbench results drops to 10MB/s or even less!
> >
> > Taking into account that mounting happens with the same options on a client side in both cases, it there some server-side trick with options?
> >
> > Time is synchronised between hosts.
> >
> > Anton.
>


This is all due to client-side effects, so there's little you can do
server-side to affect this.

lookupcache= option probably won't make a lot of difference here, but
actimeo= option likely would. actimeo= means that the client never
caches attributes, so every time it needs to check cache coherency it
has to issue a GETATTR to the server. Dialing this to a more reasonable
value (e.g. actimeo=1 or so) should still give you pretty tight cache
coherency w/o killing performance too badly.

The cache coherency logic in the Linux NFS client is pretty complex,
but typically when you have a file that's changing rapidly it should
quickly dial down the attribute cache timeout to the default minimum
(3s). None of that matters though unless you have the "writing" client
aggressively flushing the dirty data out to the server.

--
Jeff Layton <jlayton@xxxxxxxxxx>


Then why results depend on the server?
the same client mount options results in quite different results with linux and solaris servers.
there is clearly must be something.

Anton.--
To unsubscribe from this list: send the line "unsubscribe linux-nfs" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux Filesystem Development]     [Linux USB Development]     [Linux Media Development]     [Video for Linux]     [Linux NILFS]     [Linux Audio Users]     [Yosemite Info]     [Linux SCSI]

  Powered by Linux