Hi, Recently, I use glusterfs(3.2.5) server as xenserver(6.0) vms backend, the details comes behind: The linux system is centso-6.2-x86_64, I mount two 1T-disk in /gluster01 and /gluster02 in all 3 mechines. I create a distribute volume named xen_bk_vol: gluster> volume info xen_bk_vol Volume Name: xen_bk_vol Type: Distribute Status: Stopped Number of Bricks: 6 Transport-type: tcp Bricks: Brick1: 10.52.10.5:/gluster01/xen_bk_vol Brick2: 10.52.10.5:/gluster02/xen_bk_vol Brick3: 10.52.10.6:/gluster01/xen_bk_vol Brick4: 10.52.10.6:/gluster02/xen_bk_vol Brick5: 10.52.10.7:/gluster01/xen_bk_vol Brick6: 10.52.10.7:/gluster02/xen_bk_vol Options Reconfigured: network.ping-timeout: 5 auth.allow: 10.*,172.27.* gluster> I mount the volume in xenserver using : "mount 10.52.10.6:xen_bk_vol /backup -t nfs -o proto=tcp,vers=3" (My xenservers also are in the same subnet) When I use "xe vm-export" (export vm to one single xva file in xenserver) to export vms, I found something confused. When the exported vm files exceed a single point, such as 100G or 200G(I don't known the exact number), the mount point could not write data anymore and the glusterfs server (the one I used to as mount point in xenserver, as 10.52.10.6) reboots very often. I cannot get information in the glusterfs logs. Is my setting wrong or it is a bug? Thanks! -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://gluster.org/pipermail/gluster-users/attachments/20120219/28d939a6/attachment.htm>