On 2011-07-06 23:58, Papp Tamas wrote: > hi! > > I'm almost absolutely new to glusterfs. > > Until now we used Storage Platform (3.0.5). > Today we installed Ubuntu 11.04 and glusterfs 3.2.1. > > $ cat w-vol-fuse.vol > volume w-vol-client-0 > type protocol/client > option remote-host gl0 > option remote-subvolume /mnt/brick1 > option transport-type tcp > end-volume > > volume w-vol-client-1 > type protocol/client > option remote-host gl1 > option remote-subvolume /mnt/brick1 > option transport-type tcp > end-volume > > volume w-vol-client-2 > type protocol/client > option remote-host gl2 > option remote-subvolume /mnt/brick1 > option transport-type tcp > end-volume > > volume w-vol-client-3 > type protocol/client > option remote-host gl3 > option remote-subvolume /mnt/brick1 > option transport-type tcp > end-volume > > volume w-vol-client-4 > type protocol/client > option remote-host gl4 > option remote-subvolume /mnt/brick1 > option transport-type tcp > end-volume > > volume w-vol-dht > type cluster/distribute > subvolumes w-vol-client-0 w-vol-client-1 w-vol-client-2 > w-vol-client-3 w-vol-client-4 > end-volume > > volume w-vol-write-behind > type performance/write-behind > option cache-size 4MB > subvolumes w-vol-dht > end-volume > > volume w-vol-read-ahead > type performance/read-ahead > subvolumes w-vol-write-behind > end-volume > > volume w-vol-io-cache > type performance/io-cache > option cache-size 128MB > subvolumes w-vol-read-ahead > end-volume > > volume w-vol-quick-read > type performance/quick-read > option cache-size 128MB > subvolumes w-vol-io-cache > end-volume > > volume w-vol-stat-prefetch > type performance/stat-prefetch > subvolumes w-vol-quick-read > end-volume > > volume w-vol > type debug/io-stats > option latency-measurement off > option count-fop-hits off > subvolumes w-vol-stat-prefetch > end-volume > > > $ cat w-vol.gl0.mnt-brick1.vol > volume w-vol-posix > type storage/posix > option directory /mnt/brick1 > end-volume > > volume w-vol-access-control > type features/access-control > subvolumes w-vol-posix > end-volume > > volume w-vol-locks > type features/locks > subvolumes w-vol-access-control > end-volume > > volume w-vol-io-threads > type performance/io-threads > subvolumes w-vol-locks > end-volume > > volume w-vol-marker > type features/marker > option volume-uuid ad362448-7ef0-49ae-b13c-74cb82ce9be5 > option timestamp-file /etc/glusterd/vols/w-vol/marker.tstamp > option xtime off > option quota off > subvolumes w-vol-io-threads > end-volume > > volume /mnt/brick1 > type debug/io-stats > option latency-measurement off > option count-fop-hits off > subvolumes w-vol-marker > end-volume > > volume w-vol-server > type protocol/server > option transport-type tcp > option auth.addr./mnt/brick1.allow * > subvolumes /mnt/brick1 > end-volume > > > There is 5 nodes. 3 have 8 disks in RAID 6 (supermicro server, are > controller), 2 have 8 disks in raid5+spare (DL180). > Filesystem of datas was created via this command (of course a bit > different on HPs): > > mkfs.xfs -b size=4096 -d sunit=256,swidth=1536 -L gluster /dev/sda4 > > > The performance is far away that was before. I tried to modify > > performance.write-behind-window-size 4MB > gluster volume set w-vol performance.cache-size 128MB > gluster volume set w-vol nfs.disable on > > echo 512 > /sys/block/sda/queue/nr_requests > blockdev --setra 16384 /dev/sda > sysctl -w vm.swappiness=5 > sysctl -w vm.dirty_background_ratio=3 > sysctl -w vm.dirty_ratio=40 > sysctl -w kernel.sysrq=0 > > Nothing really helped. > Can somebody give some instructions? Some more information. On the node: $ dd if=/dev/zero of=adsdfgrr bs=128K count=100k oflag=direct 102400+0 records in 102400+0 records out 13421772800 bytes (13 GB) copied, 27.4022 s, 490 MB/s The same on the cluster volume is ~50-60 MB/s. Network layer is GE, nodes are connected with two NICs in bonding. I am absolutely desparated. Is it Ubuntu? Would be better with Fedora? Or does the Storage Platform run on an optimized kernel or something like that? Thank you, tamas