I have two CentOS-6 servers running version 3.12.14 of gluster-server. Each server as one brick and they are configured to replicate between the two bricks. I also have two CentOS-6 servers running version 3.12.2-18 of glusterfs. These servers use a separate VLAN. Each server has two bonded 1 Gbps NICs to communicate the gluster traffic. File transfer speeds between these servers using rsync approaches 100 MBps. The client servers mount the gluster volume using this fstab entry: 192.168.40.30:gv0 /store glusterfs defaults, attribute-timeout=600,entry-timeout=600,negative-timeout=600,fopen-keep-cache,use-readdirp=no,log-level=WARNING 1 2 Reading data from the servers to the clients is similar to the rsync speed. The problem is that writing from the clients to the mounted gluster volume is less than 8 MB/s and fluctuates from less than 500 kB/s to 8 MB/s, as measured by the pv command. Using rsync, the speed fluctuates between 2 and 5 MBps. When the bonded nics on one of the gluster servers is shut down, the write speed to the remaining online brick is now similar to the read speed I can only assume that there is something wrong in my configuration, since a greater than 10-fold decrease in write speed when the bricks are replicating makes for an unusable system. Does anyone have any ideas what the problem may be? Server volume configuration: > sudo gluster volume info Volume Name: gv0 Type: Replicate Volume ID: d96bbb99-f264-4655-95ff-f9f05ca9ff55 Status: Started Snapshot Count: 0 Number of Bricks: 1 x 2 = 2 Transport-type: tcp Bricks: Brick1: 192.168.40.20:/export/scsi/brick Brick2: 192.168.40.30:/export/scsi/brick Options Reconfigured: performance.cache-size: 1GB performance.readdir-ahead: on features.cache-invalidation: on features.cache-invalidation-timeout: 600 performance.stat-prefetch: on performance.cache-samba-metadata: on performance.cache-invalidation: on performance.md-cache-timeout: 600 network.inode-lru-limit: 250000 performance.cache-refresh-timeout: 60 performance.read-ahead: disable performance.parallel-readdir: on performance.write-behind-window-size: 4MB performance.io-thread-count: 64 performance.client-io-threads: on performance.quick-read: on performance.flush-behind: on performance.write-behind: on nfs.disable: on client.event-threads: 3 server.event-threads: 3 server.allow-insecure: on _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users