Logs are attached For claryfication: vmhost1-cluster1 -> Brick 1 vmhost2-cluster2 -> Brick 2 entrance -> Peer Time of testing (31.01.2016 16:13) Thanks for your help Regards, Dominique Werde Teil des modernen Arbeitens im Glarnerland auf www.digitalglarus.ch! Lese Neuigkeiten auf Twitter: www.twitter.com/DigitalGlarus Diskutiere mit auf Facebook: www.facebook.com/digitalglarus On 02/08/2016 04:40 PM, FNU Raghavendra Manjunath wrote: > + Pranith > > In the meantime, can you please provide the logs of all the gluster > server machines and the client machines? > > Logs can be found in /var/log/glusterfs directory. > > Regards, > Raghavendra > > On Mon, Feb 8, 2016 at 9:20 AM, Dominique Roux > <dominique.roux@xxxxxxxxxxx <mailto:dominique.roux@xxxxxxxxxxx>> wrote: > > Hi guys, > > I faced a problem a week ago. > In our environment we have three servers in a quorum. The gluster volume > is spreaded over two bricks and has the type replicated. > > We now, for simulating a fail of one brick, isolated one of the two > bricks with iptables, so that communication to the other two peers > wasn't possible anymore. > After that VMs (opennebula) which had I/O in this time crashed. > We stopped the glusterfsd hard (kill -9) and restarted it, what made > things work again (Certainly we also had to restart the failed VMs). But > I think this shouldn't happen. Since quorum was not reached (2/3 hosts > were still up and connected). > > Here some infos of our system: > OS: CentOS Linux release 7.1.1503 > Glusterfs version: glusterfs 3.7.3 > > gluster volume info: > > Volume Name: cluster1 > Type: Replicate > Volume ID: > Status: Started > Number of Bricks: 1 x 2 = 2 > Transport-type: tcp > Bricks: > Brick1: srv01:/home/gluster > Brick2: srv02:/home/gluster > Options Reconfigured: > cluster.self-heal-daemon: enable > cluster.server-quorum-type: server > network.remote-dio: enable > cluster.eager-lock: enable > performance.stat-prefetch: on > performance.io-cache: off > performance.read-ahead: off > performance.quick-read: off > server.allow-insecure: on > nfs.disable: 1 > > Hope you can help us. > > Thanks a lot. > > Best regards > Dominique > _______________________________________________ > Gluster-users mailing list > Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx> > http://www.gluster.org/mailman/listinfo/gluster-users > >
Attachment:
logs_glusterlist.tar.xz
Description: application/xz
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users