Anand Avati wrote: > Please post your spec files and the commands used to benchmark Alright, just a sec, as I'm running another benchmark with write-back's aggregate-size set to 4MB (which seems to be the maximum)... Alright, that made a bit of an improvement, but it's still slower than 1.3.10 with aggregate-size set to 1MB on the server side, and also still a *lot* slower than the write speeds of the systems involved (~70MB/sec on server1, ~35MB/sec on server2): 1.3.10: write=20.8MB/sec, read=36.1MB/sec, rewrite=12.5MB/sec 1.4pre5: write=19.3MB/sec, read=61.48MB/sec, rewrite 10.3MB/sec Server config (on both machines) is: -- volume unify-ds-brick type storage/posix option directory /srv/gluster/unify-ds end-volume volume unify-ds-lock type features/posix-locks subvolumes unify-ds-brick end-volume volume unify-ds type performance/io-threads option thread-count 2 # <= # logical CPUs option cache-size 64MB subvolumes unify-ds-lock end-volume volume server type protocol/server option transport-type tcp/server # For TCP/IP transport subvolumes unify-ds unify-ns option auth.addr.unify-ds.allow server1ip,server2ip,127.0.0.1 option auth.addr.unify-ns.allow server1ip,server2ip,127.0.0.1 end-volume -- Client config (on server1) is: -- volume mirror1-brick type protocol/client option transport-type tcp/client # for TCP/IP transport option remote-host 127.0.0.1 option remote-subvolume unify-ds # name of the remote volume end-volume volume mirror2-brick type protocol/client option transport-type tcp/client # for TCP/IP transport option remote-host 69.90.194.208 option remote-subvolume unify-ds # name of the remote volume end-volume ### READ AHEAD ### volume mirror1-read type performance/read-ahead option page-size 256kB option page-count 4 subvolumes mirror1-brick end-volume volume mirror2-read type performance/read-ahead option page-size 256kB option page-count 4 subvolumes mirror2-brick end-volume ### WRITE BEHIND ### volume mirror1 type performance/write-behind option aggregate-size 4MB option flush-behind off subvolumes mirror1-read end-volume volume mirror2 type performance/write-behind option aggregate-size 4MB option flush-behind off subvolumes mirror2-read end-volume ### TOP LEVEL VOLUMES ### volume mirror type cluster/afr subvolumes mirror1 mirror2 option read-subvolume mirror1 end-volume -- And the benchmark I'm running on server1 (after mounting on /srv/mnt) is: bonnie -s 8G -u nobody -d /srv/mnt/bonnie -f G