On 07/21/2015 08:55 AM, Punit Dambiwal wrote: > Hi Atin, > > Yes..Now i reproduce the problem again...with only 3 bricks it's working > fine... > > [image: Inline image 1] > [image: Inline image 2] > > Now when i added more 6 bricks in to the same volume..bricks added > successfully,but after adding the bricks the type of the volume has been > chaned from replicate to distributed replicate... This is expected. Earlier you had 3 bricks and since you used replica count as 3, it was a pure replicated volume. Now having another 6 bricks added to the volume the configuration becomes 3 X 3 which means distribution logic (DHT) can put in data in any of these three set of bricks, thats where we call it a distributed replicated volume. I am still not clear about your problem though. When you say "quorum is failing" what do you mean by that? A detailed description of steps performed along with etc-*-glusterd*.log file will help us analyzing the issue. > > [image: Inline image 3] > > Is there any body can help me in this...now it's more and more complex.. > > Thanks, > Punit > > On Mon, Jul 20, 2015 at 5:12 PM, Ramesh Nachimuthu <rnachimu@xxxxxxxxxx> > wrote: > >> >> >> On 07/20/2015 12:44 PM, Punit Dambiwal wrote: >> >> Hi Atin, >> >> Earlier i was adding the brick and create the volume through ovirt >> itself....this time i did it through commandline...i added all 15 bricks >> 5*3 through command line.... >> >> It seems before optimize for virt storage it all ok...but after that one >> quorum start and it failed the whole... >> >> >> "Optimize for Virt" action in oVirt sets the following volume options. >> >> group = > virt >> storage.owner-uid => 36 >> storage.owner-gid => 36 >> server.allow-insecure => on >> network.ping-timeout => 10 >> >> Regards, >> Ramesh >> >> On Mon, Jul 20, 2015 at 3:07 PM, Atin Mukherjee <amukherj@xxxxxxxxxx> >> wrote: >> >>> In one of your earlier email you mentioned that after adding a brick >>> volume status stopped working. Can you point me to the glusterd log for >>> that transaction? >>> >>> ~Atin >>> >>> On 07/20/2015 12:11 PM, Punit Dambiwal wrote: >>>> Hi Atin, >>>> >>>> Please find the below details :- >>>> >>>> [image: Inline image 1] >>>> >>>> [image: Inline image 2] >>>> >>>> Now when i set the optimize for the virt storage under ovirt and restart >>>> glusterd service on any node...it start failing the quorum.. >>>> >>>> [image: Inline image 3] >>>> >>>> [image: Inline image 4] >>>> >>>> Thanks, >>>> Punit >>>> >>>> On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit@xxxxxxxxx> >>> wrote: >>>> >>>>> HI Atin, >>>>> >>>>> Apologies for the delay response... >>>>> >>>>> 1. When you added the brick was the command successful? >>>>>>> Yes..it was successful.. >>>>> 2. If volume status is failing what's output its throwing in the >>> console >>>>> and how about the glusterd log? >>>>>>> I will reproduce the issue again and update you.. >>>>> >>>>> On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj@xxxxxxxxxx> >>>>> wrote: >>>>> >>>>>> >>>>>> >>>>>> On 07/13/2015 05:19 AM, Punit Dambiwal wrote: >>>>>>> Hi Sathees, >>>>>>> >>>>>>> With 3 bricks i can get the gluster volume status....but after added >>>>>> more >>>>>>> bricks....can not get gluster volume status.... >>>>>> The information is still incomplete in respect to analyze the problem. >>>>>> Further questions: >>>>>> >>>>>> 1. When you added the brick was the command successful? >>>>>> 2. If volume status is failing what's output its throwing in the >>> console >>>>>> and how about the glusterd log? >>>>>> >>>>>> ~Atin >>>>>>> >>>>>>> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar@xxxxxxxxxx> >>> >>> wrote: >>>>>>> >>>>>>>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote: >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> I have deployed one replica 3 storage...but i am facing some issue >>>>>> with >>>>>>>>>> quorum... >>>>>>>>>> >>>>>>>>>> Let me elaborate more :- >>>>>>>>>> >>>>>>>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No >>>>>>>>>> RAID...Just JBOD... >>>>>>>>>> 2. Gluster working fine when just add 3 HDD as below :- >>>>>>>>>> >>>>>>>>>> B HDD from server 1 >>>>>>>>>> B HDD from server 2 >>>>>>>>>> B HDD from server 3 >>>>>>>>>> >>>>>>>>>> But when i add more bricks as below :- >>>>>>>>>> >>>>>>>>>> ----------------------- >>>>>>>>>> [root@stor1 ~]# gluster volume info >>>>>>>>>> >>>>>>>>>> Volume Name: 3TB >>>>>>>>>> Type: Distributed-Replicate >>>>>>>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8 >>>>>>>>>> Status: Stopped >>>>>>>>>> Number of Bricks: 5 x 3 = 15 >>>>>>>>>> Transport-type: tcp >>>>>>>>>> Bricks: >>>>>>>>>> Brick1: stor1:/bricks/b/vol1 >>>>>>>>>> Brick2: stor2:/bricks/b/vol1 >>>>>>>>>> Brick3: stor3:/bricks/b/vol1 >>>>>>>>>> Brick4: stor1:/bricks/c/vol1 >>>>>>>>>> Brick5: stor2:/bricks/c/vol1 >>>>>>>>>> Brick6: stor3:/bricks/c/vol1 >>>>>>>>>> Brick7: stor1:/bricks/d/vol1 >>>>>>>>>> Brick8: stor2:/bricks/d/vol1 >>>>>>>>>> Brick9: stor3:/bricks/d/vol1 >>>>>>>>>> Brick10: stor1:/bricks/e/vol1 >>>>>>>>>> Brick11: stor2:/bricks/e/vol1 >>>>>>>>>> Brick12: stor3:/bricks/e/vol1 >>>>>>>>>> Brick13: stor1:/bricks/f/vol1 >>>>>>>>>> Brick14: stor2:/bricks/f/vol1 >>>>>>>>>> Brick15: stor3:/bricks/f/vol1 >>>>>>>>>> Options Reconfigured: >>>>>>>>>> nfs.disable: off >>>>>>>>>> user.cifs: enable >>>>>>>>>> auth.allow: * >>>>>>>>>> performance.quick-read: off >>>>>>>>>> performance.read-ahead: off >>>>>>>>>> performance.io-cache: off >>>>>>>>>> performance.stat-prefetch: off >>>>>>>>>> cluster.eager-lock: enable >>>>>>>>>> network.remote-dio: enable >>>>>>>>>> cluster.quorum-type: auto >>>>>>>>>> cluster.server-quorum-type: server >>>>>>>>>> storage.owner-uid: 36 >>>>>>>>>> storage.owner-gid: 36 >>>>>>>>>> -------------------------------- >>>>>>>>>> >>>>>>>>>> Brick added successfully without any error but after 1 min quorum >>>>>> failed >>>>>>>>>> and gluster stop working... >>>>>>>>>> >>>>>>>>> Punit, >>>>>>>> >>>>>>>> And what do you mean by quorum failed ? >>>>>>>> What is effect that you are seeing ? >>>>>>>> Could you provide output of 'gluster volume status' as well ? >>>>>>>> >>>>>>>> -- Sathees >>>>>>>> >>>>>>>> >>>>>>>> What do log files say? >>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> Punit >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Gluster-users mailing list >>> >>>>>>> <Gluster-users@xxxxxxxxxxx>Gluster-users@xxxxxxxxxxx >>>>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users >>>>>>>>>> >>>>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>>> -- >>>>>> ~Atin >>>>>> >>>>> >>>>> >>>> >>> >>> -- >>> ~Atin >>> >> >> >> >> _______________________________________________ >> Gluster-users mailing listGluster-users@gluster.orghttp://www.gluster.org/mailman/listinfo/gluster-users >> >> >> > -- ~Atin _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users