> I'm using gluster 3.3.0 and 3.3.1 with xfs bricks and kvm based VM's > using qcow2 files on gluster volume fuse mounts. CentOS6.2 through 6.4 > w/CloudStack 3.0.2 - 4.1.0. > > I've not had any problems. Here is 1 host in a small 3 host cluster > (using the cloudstack terminology). about 30 VM's are running across > these 3 hosts - which all contribute to the volume with 2 bricks each. > I'll also attach a virsh dumpxml for you to take a look at. > > [root ~]# w > 06:21:53 up 320 days, 7:23, 1 user, load average: 1.41, 1.07, 0.79 > USER TTY FROM LOGIN@ IDLE JCPU PCPU WHAT > root pts/9 10.100.0.100 06:21 0.00s 0.00s 0.00s w > [root ~]# cat /etc/redhat-release > CentOS release 6.3 (Final) > [root ~]# rpm -qa | grep gluster > glusterfs-server-3.3.0-1.el6.x86_64 > glusterfs-fuse-3.3.0-1.el6.x86_64 > glusterfs-3.3.0-1.el6.x86_64 > [root@ ~]# cat /etc/fstab | grep glust > /dev/storage/glust0 /gluster/0 xfs defaults,inode64 0 0 > /dev/storage/glust1 /gluster/1 xfs defaults,inode64 0 0 > 172.16.0.11:qcow2-share /gluster/qcow2 glusterfs defaults,_netdev 0 0 > [root at cs0.la.vorstack.net ~]# df -h > [cut.....] > /dev/mapper/storage-glust0 > 2.0T 217G 1.8T 11% /gluster/0 > /dev/mapper/storage-glust1 > 2.0T 148G 1.9T 8% /gluster/1 > 172.16.0.11:qcow2-share > 6.0T 472G 5.6T 8% /gluster/qcow2 > [root@ ~]# virsh list > Id Name State > ---------------------------------------------------- > 10 i-2-19-VM running > 21 i-3-44-VM running > 22 i-2-12-VM running > 28 i-4-58-VM running > 37 s-5-VM running > 38 v-2-VM running > 39 i-2-56-VM running > 41 i-7-59-VM running > 46 i-4-87-VM running > [root@ ~]# gluster volume info > > Volume Name: qcow2-share > Type: Distributed-Replicate > Volume ID: 22fcbaa9-4b2d-4d84-9353-eb77abcaf0db > Status: Started > Number of Bricks: 3 x 2 = 6 > Transport-type: tcp > Bricks: > Brick1: 172.16.0.10:/gluster/0 > Brick2: 172.16.0.11:/gluster/0 > Brick3: 172.16.0.12:/gluster/0 > Brick4: 172.16.0.10:/gluster/1 > Brick5: 172.16.0.11:/gluster/1 > Brick6: 172.16.0.12:/gluster/1 > [root@ ~]# gluster volume status > Status of volume: qcow2-share > Gluster process Port Online Pid > ------------------------------------------------------------------------------ > Brick 172.16.0.10:/gluster/0 24009 Y 1873 > Brick 172.16.0.11:/gluster/0 24009 Y 1831 > Brick 172.16.0.12:/gluster/0 24009 Y 1938 > Brick 172.16.0.10:/gluster/1 24010 Y 1878 > Brick 172.16.0.11:/gluster/1 24010 Y 1837 > Brick 172.16.0.12:/gluster/1 24010 Y 1953 > NFS Server on localhost 38467 Y 1899 > Self-heal Daemon on localhost N/A Y 1909 > NFS Server on 172.16.0.12 38467 Y 1959 > Self-heal Daemon on 172.16.0.12 N/A Y 1964 > NFS Server on 172.16.0.11 38467 Y 1843 > Self-heal Daemon on 172.16.0.11 N/A Y 1848 > > This information (including the attached xml) is very helpful, thank you! Can you provide the xml of your (KVM) gluster storage pool: "virsh pool-dumpxml <pool>" Have you ever tried using "raw" virtio disk images? When I try to use qcow2, I get errors when trying to start my VM: qemu-kvm: -drive file=/var/lib/libvirt/images/xfs/kvm2.img,if=none,id=drive-virtio-disk1,format=qcow2,cache=none: could not open disk image /var/lib/libvirt/images/xfs/kvm2.img: Invalid argument -Jacob