Hi. I wanted to update with results from my tests. So, up to full disk it worked fine, no memory leaks and equal size was reported on all the directories - export on both servers, and the mounted one on client. I didn't stop/start the cluster nodes - just let it run without any stop. After it reached the full size of the cluster, things seem to break somewhat: DF reports one size on client and on one of the servers, but another size on the another server: client: glusterfs 31G 29G 1.3G 96% /mnt/media server 1: /dev/hda4 31G 29G 1.3G 96% /media server 2: /dev/hda4 31G 18G 13G 60% /media DU over mounted and exported directories reports same sizes on all the nodes: 18G So, because one of the servers reports that almost all space is taken while it's not true, the cluster has lost almost half of the available space. AFAIK GlusterFS always reports the free space provided by the first server indeed, but it completely incorrect here. What is more interesting, that it always happens with the first server defined in client vol file. I switched the servers order this time just to verify it, and it happened on the server that become the first one. There are no 0 size files this time though. So, what should be my steps to help the devs to find this issue? I'm currently leaving all running for some time to see how it develops further. Regards. -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://zresearch.com/pipermail/gluster-users/attachments/20090402/54fe83da/attachment-0001.htm>