Hi, Same problem on 3.8.1. Even on loopback interface (traffic not leaves gluster node): Writing locally to replica 2 volume (each brick is separate local RAID6): 613 MB/sec Writing locally to 1-brick volume: 877 MB/sec Writing locally to the brick itself (directly to XFS): 1400 MB/sec Tests were performed using fio with following settings: bs=4096k ioengine=libaio iodepth=32 direct=0 runtime=600 directory=/R1 numjobs=1 rw=write size=40g Even with direct=1 the brick itself gives 1400 MB/sec. 1-brick volume profiling below: # gluster volume profile test-data-03 info Brick: gluster-01:/R1/test-data-03 ----------------------------------------------- Cumulative Stats: Block Size: 131072b+ 262144b+ No. of Reads: 0 0 No. of Writes: 889072 20 %-latency Avg-latency Min-Latency Max-Latency No. of calls Fop --------- ----------- ----------- ----------- ------------ ---- 0.00 0.00 us 0.00 us 0.00 us 3 RELEASE 100.00 122.96 us 67.00 us 42493.00 us 208598 WRITE Duration: 1605 seconds Data Read: 0 bytes Data Written: 116537688064 bytes Interval 0 Stats: Block Size: 131072b+ 262144b+ No. of Reads: 0 0 No. of Writes: 889072 20 %-latency Avg-latency Min-Latency Max-Latency No. of calls Fop --------- ----------- ----------- ----------- ------------ ---- 0.00 0.00 us 0.00 us 0.00 us 3 RELEASE 100.00 122.96 us 67.00 us 42493.00 us 208598 WRITE Duration: 1605 seconds Data Read: 0 bytes Data Written: 116537688064 bytes # As you can see all writes are performed using 128 KB block size. And it looks like a bottleneck. Which was discussed previously btw: http://www.gluster.org/pipermail/gluster-devel/2013-March/038821.html Using GFAPI to access the volume shows better speed, but still far from raw brick. fio tests with ioengine=gfapi gives following: Writing locally to replica 2 volume (each brick is separate local RAID6): 680 MB/sec Writing locally to 1-brick volume: 960 MB/sec Accorging to 1-brick volume profile 128 KB blocks no more used: # gluster volume profile tzk-data-03 info Brick: j-gluster-01.vcod.jet.su:/R1/tzk-data-03 ----------------------------------------------- Cumulative Stats: Block Size: 4194304b+ No. of Reads: 0 No. of Writes: 9211 %-latency Avg-latency Min-Latency Max-Latency No. of calls Fop --------- ----------- ----------- ----------- ------------ ---- 100.00 2237.67 us 1880.00 us 5785.00 us 8701 WRITE Duration: 49 seconds Data Read: 0 bytes Data Written: 38633734144 bytes Interval 0 Stats: Block Size: 4194304b+ No. of Reads: 0 No. of Writes: 9211 %-latency Avg-latency Min-Latency Max-Latency No. of calls Fop --------- ----------- ----------- ----------- ------------ ---- 100.00 2237.67 us 1880.00 us 5785.00 us 8701 WRITE Duration: 49 seconds Data Read: 0 bytes Data Written: 38633734144 bytes [root@j-gluster-01 ~]# So, it may be worth to try using NFS Ganesha with GFAPI plugin. > 3 авг. 2016 г., в 9:40, Kaamesh Kamalaaharan <kaamesh@xxxxxxxxxxxxx> написал(а): > > Hi , > I have gluster 3.6.2 installed on my server network. Due to internal issues we are not allowed to upgrade the gluster version. All the clients are on the same version of gluster. When transferring files to/from the clients or between my nodes over the 10gb network, the transfer rate is capped at 450Mb/s .Is there any way to increase the transfer speeds for gluster mounts? > > Our server setup is as following: > > 2 gluster servers -gfs1 and gfs2 > volume name : gfsvolume > 3 clients - hpc1, hpc2,hpc3 > gluster volume mounted on /export/gfsmount/ > > > > The following is the average results what i did so far: > > 1) test bandwith with iperf between all machines - 9.4 GiB/s > 2) test write speed with dd > dd if=/dev/zero of=/export/gfsmount/testfile bs=1G count=1 > > result=399Mb/s > > 3) test read speed with dd > dd if=/export/gfsmount/testfile of=/dev/zero bs=1G count=1 > > result=284MB/s > > My gluster volume configuration: > > Volume Name: gfsvolume > Type: Replicate > Volume ID: a29bd2fb-b1ef-4481-be10-c2f4faf4059b > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: gfs1:/export/sda/brick > Brick2: gfs2:/export/sda/brick > Options Reconfigured: > performance.quick-read: off > network.ping-timeout: 30 > network.frame-timeout: 90 > performance.cache-max-file-size: 2MB > cluster.server-quorum-type: none > nfs.addr-namelookup: off > nfs.trusted-write: off > performance.write-behind-window-size: 4MB > cluster.data-self-heal-algorithm: diff > performance.cache-refresh-timeout: 60 > performance.cache-size: 1GB > cluster.quorum-type: fixed > auth.allow: 172.* > cluster.quorum-count: 1 > diagnostics.latency-measurement: on > diagnostics.count-fop-hits: on > cluster.server-quorum-ratio: 50% > > Any help would be appreciated. > Thanks, > Kaamesh > > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://www.gluster.org/mailman/listinfo/gluster-users -- Dmitry Glushenok Jet Infosystems _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users