Hi Anand, This is one entry (of many) in the client log when bringing my second node of the cluster back up, the glustershd.log is completely silent at this point. If your interested in seeing the nodes split & reconnect, the relevant glustershd.log section is at http://pastebin.com/0Va3RxDD many thanks! Jake > Was this the client log or the glustershd log? > > Thanks, > Avati > > On Mon, Jul 9, 2012 at 8:23 AM, Jake Grimmett <jog at mrc-lmb.cam.ac.uk> > wrote: > >> Hi Fernando / Christian, >> >> Many thanks for getting back to me. >> >> Slow writes are acceptable; most of our VM's are small web servers with >> low traffic. My aim is to have a fully self-contained two server KVM >> cluster with live migration, no external storage and the ability to >> reboot >> either node with zero VM downtime. We seem to be "almost there", bar a >> hiccup when the self-heal is in progress and some minor grumbles from >> sanlock (which might be fixed by the new sanlock in RHEL 6.3) >> >> Incidentally, the logs shows a "diff" self heal on a node reboot: >> >> [2012-07-09 16:04:06.743512] I >> [afr-self-heal-algorithm.c:**122:sh_loop_driver_done] >> 0-gluster-rep-replicate-0: diff self-heal on /box1-clone2.img: >> completed. >> (16 blocks of 16974 were different (0.09%)) >> >> So, does this log show "Granular locking" occurring, or does it just >> happen transparently when a file exceeds a certain size? >> >> many thanks >> >> Jake >> >> >> >> On 07/09/2012 04:01 PM, Fernando Frediani (Qube) wrote: >> >>> Jake, >>> >>> I haven?t had a chanced to test with my KVM cluster yet but it should >>> be >>> a default things from 3.3. >>> >>> Just be in mind that running Virtual Machines is NOT a supported things >>> for Redhat Storage server according to Redhat Sales people. They said >>> towards the end of the year. As you might have observed performance >>> specially for write isn?t any near fantastic. >>> >>> >>> Fernando >>> >>> *From:*gluster-users-bounces@**gluster.org<gluster-users-bounces at gluster.org> >>> [mailto:gluster-users-bounces@**gluster.org<gluster-users-bounces at gluster.org>] >>> *On Behalf Of *Christian Wittwer >>> *Sent:* 09 July 2012 15:51 >>> *To:* Jake Grimmett >>> *Cc:* gluster-users at gluster.org >>> *Subject:* Re: "Granular locking" - does this need to >>> be >>> >>> enabled in 3.3.0 ? >>> >>> Hi Jake >>> >>> I can confirm exact the same behaviour with gluster 3.3.0 on Ubuntu >>> 12.04. During the self-heal process the VM gets 100% I/O wait and is >>> locked. >>> >>> After the self-heal the root filesystem was read-only which forced me >>> to >>> do a reboot and fsck. >>> >>> Cheers, >>> >>> Christian >>> >>> 2012/7/9 Jake Grimmett <jog at mrc-lmb.cam.ac.uk >>> <mailto:jog at mrc-lmb.cam.ac.uk>**> >>> >>> >>> Dear All, >>> >>> I have a pair of Scientific Linux 6.2 servers, acting as KVM >>> virtualisation hosts for ~30 VM's. The VM images are stored in a >>> replicated gluster volume shared between the two servers. Live >>> migration >>> works fine, and the sanlock prevents me from (stupidly) starting the >>> same VM on both machines. Each server has 10GB ethernet and a 10 disk >>> RAID5 array. >>> >>> If I migrate all the VM's to server #1 and shutdown server #2, all >>> works >>> perfectly with no interruption. When I restart server #2, the VM's >>> freeze while the self-heal process is running - and this healing can >>> take a long time. >>> >>> I'm not sure if "Granular Locking" is on. It's listed as a "technology >>> preview" in the Redhat Storage server 2 notes - do I need to do >>> anything >>> to enable it? >>> >>> i.e. set "cluster.data-self-heal-**algorithm" to diff ? >>> or edit "cluster.self-heal-window-**size" ? >>> >>> any tips from other people doing similar much appreciated! >>> >>> Many thanks, >>> >>> Jake >>> >>> jog <---at---> mrc-lmb.cam.ac.uk <http://mrc-lmb.cam.ac.uk> >>> ______________________________**_________________ >>> Gluster-users mailing list >>> Gluster-users at gluster.org >>> <mailto:Gluster-users at gluster.**org<Gluster-users at gluster.org> >>> > >>> http://gluster.org/cgi-bin/**mailman/listinfo/gluster-users<http://gluster.org/cgi-bin/mailman/listinfo/gluster-users> >>> >>> >> >> -- >> Dr Jake Grimmett >> Head Of Scientific Computing >> MRC Laboratory of Molecular Biology >> Hills Road, Cambridge, CB2 0QH, UK. >> Phone 01223 402219 >> Mobile 0776 9886539 >> >> ______________________________**_________________ >> Gluster-users mailing list >> Gluster-users at gluster.org >> http://gluster.org/cgi-bin/**mailman/listinfo/gluster-users<http://gluster.org/cgi-bin/mailman/listinfo/gluster-users> >> > -- Dr Jake Grimmett Head Of Scientific Computing MRC Laboratory of Molecular Biology Hills Road, Cambridge, CB2 0QH, UK. Phone 01223 402219 Mobile 0776 9886539