Re: Quorum For replica 3 storage

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



In one of your earlier email you mentioned that after adding a brick
volume status stopped working. Can you point me to the glusterd log for
that transaction?

~Atin

On 07/20/2015 12:11 PM, Punit Dambiwal wrote:
> Hi Atin,
> 
> Please find the below details :-
> 
> [image: Inline image 1]
> 
> [image: Inline image 2]
> 
> Now when i set the optimize for the virt storage under ovirt and restart
> glusterd service on any node...it start failing the quorum..
> 
> [image: Inline image 3]
> 
> [image: Inline image 4]
> 
> Thanks,
> Punit
> 
> On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit@xxxxxxxxx> wrote:
> 
>> HI Atin,
>>
>> Apologies for the delay response...
>>
>> 1. When you added the brick was the command successful?
>>>> Yes..it was successful..
>> 2. If volume status is failing what's output its throwing in the console
>> and how about the glusterd log?
>>>> I will reproduce the issue again and update you..
>>
>> On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj@xxxxxxxxxx>
>> wrote:
>>
>>>
>>>
>>> On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
>>>> Hi Sathees,
>>>>
>>>> With 3 bricks i can get the gluster volume status....but after added
>>> more
>>>> bricks....can not get gluster volume status....
>>> The information is still incomplete in respect to analyze the problem.
>>> Further questions:
>>>
>>> 1. When you added the brick was the command successful?
>>> 2. If volume status is failing what's output its throwing in the console
>>> and how about the glusterd log?
>>>
>>> ~Atin
>>>>
>>>> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar@xxxxxxxxxx>
>>> wrote:
>>>>
>>>>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
>>>>>
>>>>>>
>>>>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> I have deployed one replica 3 storage...but i am facing some issue
>>> with
>>>>>>> quorum...
>>>>>>>
>>>>>>> Let me elaborate more :-
>>>>>>>
>>>>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No
>>>>>>> RAID...Just JBOD...
>>>>>>> 2. Gluster working fine when just add 3 HDD as below :-
>>>>>>>
>>>>>>> B HDD from server 1
>>>>>>> B HDD from server 2
>>>>>>> B HDD from server 3
>>>>>>>
>>>>>>> But when i add more bricks as below :-
>>>>>>>
>>>>>>> -----------------------
>>>>>>> [root@stor1 ~]# gluster volume info
>>>>>>>
>>>>>>> Volume Name: 3TB
>>>>>>> Type: Distributed-Replicate
>>>>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
>>>>>>> Status: Stopped
>>>>>>> Number of Bricks: 5 x 3 = 15
>>>>>>> Transport-type: tcp
>>>>>>> Bricks:
>>>>>>> Brick1: stor1:/bricks/b/vol1
>>>>>>> Brick2: stor2:/bricks/b/vol1
>>>>>>> Brick3: stor3:/bricks/b/vol1
>>>>>>> Brick4: stor1:/bricks/c/vol1
>>>>>>> Brick5: stor2:/bricks/c/vol1
>>>>>>> Brick6: stor3:/bricks/c/vol1
>>>>>>> Brick7: stor1:/bricks/d/vol1
>>>>>>> Brick8: stor2:/bricks/d/vol1
>>>>>>> Brick9: stor3:/bricks/d/vol1
>>>>>>> Brick10: stor1:/bricks/e/vol1
>>>>>>> Brick11: stor2:/bricks/e/vol1
>>>>>>> Brick12: stor3:/bricks/e/vol1
>>>>>>> Brick13: stor1:/bricks/f/vol1
>>>>>>> Brick14: stor2:/bricks/f/vol1
>>>>>>> Brick15: stor3:/bricks/f/vol1
>>>>>>> Options Reconfigured:
>>>>>>> nfs.disable: off
>>>>>>> user.cifs: enable
>>>>>>> auth.allow: *
>>>>>>> performance.quick-read: off
>>>>>>> performance.read-ahead: off
>>>>>>> performance.io-cache: off
>>>>>>> performance.stat-prefetch: off
>>>>>>> cluster.eager-lock: enable
>>>>>>> network.remote-dio: enable
>>>>>>> cluster.quorum-type: auto
>>>>>>> cluster.server-quorum-type: server
>>>>>>> storage.owner-uid: 36
>>>>>>> storage.owner-gid: 36
>>>>>>> --------------------------------
>>>>>>>
>>>>>>> Brick added successfully without any error but after 1 min quorum
>>> failed
>>>>>>> and gluster stop working...
>>>>>>>
>>>>>> Punit,
>>>>>
>>>>> And what do you mean by quorum failed ?
>>>>> What is effect that you are seeing ?
>>>>> Could you provide output of 'gluster volume status' as well ?
>>>>>
>>>>> -- Sathees
>>>>>
>>>>>
>>>>>  What do log files say?
>>>>>>
>>>>>>> Thanks,
>>>>>>> Punit
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Gluster-users mailing list
>>>>>>> Gluster-users@xxxxxxxxxxx
>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>
>>> --
>>> ~Atin
>>>
>>
>>
> 

-- 
~Atin
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users



[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux