hi! I'm almost absolutely new to glusterfs. Until now we used Storage Platform (3.0.5). Today we installed Ubuntu 11.04 and glusterfs 3.2.1. $ cat w-vol-fuse.vol volume w-vol-client-0 type protocol/client option remote-host gl0 option remote-subvolume /mnt/brick1 option transport-type tcp end-volume volume w-vol-client-1 type protocol/client option remote-host gl1 option remote-subvolume /mnt/brick1 option transport-type tcp end-volume volume w-vol-client-2 type protocol/client option remote-host gl2 option remote-subvolume /mnt/brick1 option transport-type tcp end-volume volume w-vol-client-3 type protocol/client option remote-host gl3 option remote-subvolume /mnt/brick1 option transport-type tcp end-volume volume w-vol-client-4 type protocol/client option remote-host gl4 option remote-subvolume /mnt/brick1 option transport-type tcp end-volume volume w-vol-dht type cluster/distribute subvolumes w-vol-client-0 w-vol-client-1 w-vol-client-2 w-vol-client-3 w-vol-client-4 end-volume volume w-vol-write-behind type performance/write-behind option cache-size 4MB subvolumes w-vol-dht end-volume volume w-vol-read-ahead type performance/read-ahead subvolumes w-vol-write-behind end-volume volume w-vol-io-cache type performance/io-cache option cache-size 128MB subvolumes w-vol-read-ahead end-volume volume w-vol-quick-read type performance/quick-read option cache-size 128MB subvolumes w-vol-io-cache end-volume volume w-vol-stat-prefetch type performance/stat-prefetch subvolumes w-vol-quick-read end-volume volume w-vol type debug/io-stats option latency-measurement off option count-fop-hits off subvolumes w-vol-stat-prefetch end-volume $ cat w-vol.gl0.mnt-brick1.vol volume w-vol-posix type storage/posix option directory /mnt/brick1 end-volume volume w-vol-access-control type features/access-control subvolumes w-vol-posix end-volume volume w-vol-locks type features/locks subvolumes w-vol-access-control end-volume volume w-vol-io-threads type performance/io-threads subvolumes w-vol-locks end-volume volume w-vol-marker type features/marker option volume-uuid ad362448-7ef0-49ae-b13c-74cb82ce9be5 option timestamp-file /etc/glusterd/vols/w-vol/marker.tstamp option xtime off option quota off subvolumes w-vol-io-threads end-volume volume /mnt/brick1 type debug/io-stats option latency-measurement off option count-fop-hits off subvolumes w-vol-marker end-volume volume w-vol-server type protocol/server option transport-type tcp option auth.addr./mnt/brick1.allow * subvolumes /mnt/brick1 end-volume There is 5 nodes. 3 have 8 disks in RAID 6 (supermicro server, are controller), 2 have 8 disks in raid5+spare (DL180). Filesystem of datas was created via this command (of course a bit different on HPs): mkfs.xfs -b size=4096 -d sunit=256,swidth=1536 -L gluster /dev/sda4 The performance is far away that was before. I tried to modify performance.write-behind-window-size 4MB gluster volume set w-vol performance.cache-size 128MB gluster volume set w-vol nfs.disable on echo 512 > /sys/block/sda/queue/nr_requests blockdev --setra 16384 /dev/sda sysctl -w vm.swappiness=5 sysctl -w vm.dirty_background_ratio=3 sysctl -w vm.dirty_ratio=40 sysctl -w kernel.sysrq=0 Nothing really helped. Can somebody give some instructions? Thank you, tamas