On Thu, Jun 12, 2014 at 10:59:23AM +0200, Ivano Talamo wrote: > Hi Jae, > > starting the vm works fine, but only in the old-way, ie. with the > path to the mounted volume. > If trying to start with the libgfapi the virsh start commands waits > forever (is blocked on a futex). > > Maybe I should leave the libgfapi for the future. Have you taken care of setting the required glusterd and volume options? these are described here: - http://www.gluster.org/community/documentation/index.php/Libgfapi_with_qemu_libvirt After setting the server.allow-insecure volume option, a stop and a start of the volume is needed (in 3.5). HTH, Niels > > Thanks, > Ivano > > On 6/12/14 7:11 AM, Jae Park wrote: > >Ivano, > > > >Did you try to start the vm on the replicated volume? Does it work? > >I remember a vm on a replicated volume failed to start (from virsh) in > >3.5.0 due to similar errors, but now I installed 3.5.1beta2 and it > >starts > >up successfully even though qemu-img still shows the same error, which I > >can ignore now. > > > >Beta2 download: > >http://download.gluster.org/pub/gluster/glusterfs/qa-releases/3.5.1beta2/ > > > >Jae > > > >On 6/11/14 10:55 AM, "Ivano Talamo" <Ivano.Talamo@xxxxxxxxxxxxx> wrote: > > > >>Hello, > >>I recently update 2 servers (Scientific Linux 6) with a replicate volume > >>from gluster 3.4 to 3.5.0-2. > >>The volume was previously used to host qemu/kvm VM images accessed via a > >>fuse-mounted mount-point. > >>Now I would like to use the libgfapi but I'm seeing this error: > >> > >>[root@cmsrm-service02 ~]# qemu-img info > >>gluster://cmsrm-service02/vol1/vms/disks/cmsrm-ui01.raw2 > >>[2014-06-11 17:47:22.084842] E [afr-common.c:3959:afr_notify] > >>0-vol1-replicate-0: All subvolumes are down. Going offline until atleast > >>one of them comes back up. > >>image: gluster://cmsrm-service03/vol1/vms/disks/cmsrm-ui01.raw2 > >>file format: raw > >>virtual size: 20G (21474836480 bytes) > >>disk size: 4.7G > >>[2014-06-11 17:47:22.318034] E [afr-common.c:3959:afr_notify] > >>0-vol1-replicate-0: All subvolumes are down. Going offline until atleast > >>one of them comes back up. > >> > >>The error message does not appear if I access the file via the > >>mount-point. > >> > >>The volume seems fine: > >>[root@cmsrm-service02 ~]# gluster volume info > >> > >>Volume Name: vol1 > >>Type: Replicate > >>Volume ID: 35de92de-d6b3-4784-9ccb-65518e014a49 > >>Status: Started > >>Number of Bricks: 1 x 2 = 2 > >>Transport-type: tcp > >>Bricks: > >>Brick1: cmsrm-service02:/brick/vol1 > >>Brick2: cmsrm-service03:/brick/vol1 > >>Options Reconfigured: > >>server.allow-insecure: on > >>[root@cmsrm-service02 ~]# gluster volume status > >>Status of volume: vol1 > >>Gluster process Port Online Pid > >>-------------------------------------------------------------------------- > >>---- > >>Brick cmsrm-service02:/brick/vol1 49152 Y > >>16904 > >>Brick cmsrm-service03:/brick/vol1 49152 Y > >>12868 > >>NFS Server on localhost 2049 Y 4263 > >>Self-heal Daemon on localhost N/A Y 4283 > >>NFS Server on 141.108.36.8 2049 Y 13679 > >>Self-heal Daemon on 141.108.36.8 N/A Y 13691 > >> > >>Task Status of Volume vol1 > >>-------------------------------------------------------------------------- > >>---- > >>There are no active volume tasks > >> > >> > >> > >>Thank you, > >>Ivano > >> > >> > > > > > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx > http://supercolony.gluster.org/mailman/listinfo/gluster-users _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://supercolony.gluster.org/mailman/listinfo/gluster-users