On 07/13/2015 05:19 AM, Punit Dambiwal wrote: > Hi Sathees, > > With 3 bricks i can get the gluster volume status....but after added more > bricks....can not get gluster volume status.... The information is still incomplete in respect to analyze the problem. Further questions: 1. When you added the brick was the command successful? 2. If volume status is failing what's output its throwing in the console and how about the glusterd log? ~Atin > > On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar@xxxxxxxxxx> wrote: > >> On 07/11/2015 02:46 PM, Atin Mukherjee wrote: >> >>> >>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote: >>> >>>> Hi, >>>> >>>> I have deployed one replica 3 storage...but i am facing some issue with >>>> quorum... >>>> >>>> Let me elaborate more :- >>>> >>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No >>>> RAID...Just JBOD... >>>> 2. Gluster working fine when just add 3 HDD as below :- >>>> >>>> B HDD from server 1 >>>> B HDD from server 2 >>>> B HDD from server 3 >>>> >>>> But when i add more bricks as below :- >>>> >>>> ----------------------- >>>> [root@stor1 ~]# gluster volume info >>>> >>>> Volume Name: 3TB >>>> Type: Distributed-Replicate >>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8 >>>> Status: Stopped >>>> Number of Bricks: 5 x 3 = 15 >>>> Transport-type: tcp >>>> Bricks: >>>> Brick1: stor1:/bricks/b/vol1 >>>> Brick2: stor2:/bricks/b/vol1 >>>> Brick3: stor3:/bricks/b/vol1 >>>> Brick4: stor1:/bricks/c/vol1 >>>> Brick5: stor2:/bricks/c/vol1 >>>> Brick6: stor3:/bricks/c/vol1 >>>> Brick7: stor1:/bricks/d/vol1 >>>> Brick8: stor2:/bricks/d/vol1 >>>> Brick9: stor3:/bricks/d/vol1 >>>> Brick10: stor1:/bricks/e/vol1 >>>> Brick11: stor2:/bricks/e/vol1 >>>> Brick12: stor3:/bricks/e/vol1 >>>> Brick13: stor1:/bricks/f/vol1 >>>> Brick14: stor2:/bricks/f/vol1 >>>> Brick15: stor3:/bricks/f/vol1 >>>> Options Reconfigured: >>>> nfs.disable: off >>>> user.cifs: enable >>>> auth.allow: * >>>> performance.quick-read: off >>>> performance.read-ahead: off >>>> performance.io-cache: off >>>> performance.stat-prefetch: off >>>> cluster.eager-lock: enable >>>> network.remote-dio: enable >>>> cluster.quorum-type: auto >>>> cluster.server-quorum-type: server >>>> storage.owner-uid: 36 >>>> storage.owner-gid: 36 >>>> -------------------------------- >>>> >>>> Brick added successfully without any error but after 1 min quorum failed >>>> and gluster stop working... >>>> >>> Punit, >> >> And what do you mean by quorum failed ? >> What is effect that you are seeing ? >> Could you provide output of 'gluster volume status' as well ? >> >> -- Sathees >> >> >> What do log files say? >>> >>>> Thanks, >>>> Punit >>>> >>>> >>>> >>>> _______________________________________________ >>>> Gluster-users mailing list >>>> Gluster-users@xxxxxxxxxxx >>>> http://www.gluster.org/mailman/listinfo/gluster-users >>>> >>>> >> > -- ~Atin _______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users