Hi again, Addendum: I explicitly set direct-io to 'disable' and now the performance is at 110 MB/s ... which is what I would expect over a 1Gb link. thanks! Doug On 10/19/2012 07:29 AM, Doug Schouten wrote: > Hi Fernando, Anand, > > Thanks for the suggestion. I removed the enable-direct-io option in > /etc/fstab, remounted, and the performance is the same within > statistical precision. > > cheers, Doug > > > On 10/19/2012 02:13 AM, Fernando Frediani (Qube) wrote: >> Hi Doug, >> >> Try to make the change suggested by Anand and let us know how you get >> on. I am interested to hear about the performance on 3.3 because bad >> performance has been subject of many emails for a while here. >> >> Regards, >> >> Fernando >> >> -----Original Message----- >> From: gluster-users-bounces at gluster.org >> [mailto:gluster-users-bounces at gluster.org] On Behalf Of Doug Schouten >> Sent: 19 October 2012 02:45 >> To: gluster-users at gluster.org >> Subject: performance in 3.3 >> >> Hi, >> >> I am noticing a rather slow read performance using GlusterFS 3.3 with >> the following configuration: >> >> Number of Bricks: 4 >> Transport-type: tcp >> Bricks: >> Brick1: server1:/srv/data >> Brick2: server2:/srv/data >> Brick3: server3:/srv/data >> Brick4: server4:/srv/data >> Options Reconfigured: >> features.quota: off >> features.quota-timeout: 1800 >> performance.flush-behind: on >> performance.io-thread-count: 64 >> performance.quick-read: on >> performance.stat-prefetch: on >> performance.io-cache: on >> performance.write-behind: on >> performance.read-ahead: on >> performance.write-behind-window-size: 4MB >> performance.cache-refresh-timeout: 1 >> performance.cache-size: 4GB >> nfs.rpc-auth-allow: none >> network.frame-timeout: 60 >> nfs.disable: on >> performance.cache-max-file-size: 1GB >> >> >> The servers are connected with bonded 1Gb ethernet, and have LSI >> MegaRAID arrays with 12x1 TB disks in RAID-6 array, using XFS file >> system mounted like: >> >> xfs logbufs=8,logbsize=32k,noatime,nodiratime 0 0 >> >> and we use the FUSE client >> >> localhost:/global /global glusterfs >> defaults,direct-io-mode=enable,log-level=WARNING,log-file=/var/log/gluster.log >> >> 0 0 >> >> Our files are all >= 2MB. When rsync-ing we see about 50MB/s read >> performance which improves to 250MB/s after the first copy. This >> indicates to me that the disk caching is working as expected. However >> I am rather surprised by the low 50MB/s read speed; this is too low to >> be limited by network, and the native disk read performance is way >> better. >> Is there some configuration that can improve this situation? >> >> thanks, >> >> > -- Doug Schouten Research Associate TRIUMF