Re: Quorum For replica 3 storage

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



HI Atin,

Apologies for the delay response...

1. When you added the brick was the command successful?
>> Yes..it was successful..
2. If volume status is failing what's output its throwing in the console
and how about the glusterd log?
>> I will reproduce the issue again and update you..

On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:


On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
> Hi Sathees,
>
> With 3 bricks i can get the gluster volume status....but after added more
> bricks....can not get gluster volume status....
The information is still incomplete in respect to analyze the problem.
Further questions:

1. When you added the brick was the command successful?
2. If volume status is failing what's output its throwing in the console
and how about the glusterd log?

~Atin
>
> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar@xxxxxxxxxx> wrote:
>
>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
>>
>>>
>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
>>>
>>>> Hi,
>>>>
>>>> I have deployed one replica 3 storage...but i am facing some issue with
>>>> quorum...
>>>>
>>>> Let me elaborate more :-
>>>>
>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No
>>>> RAID...Just JBOD...
>>>> 2. Gluster working fine when just add 3 HDD as below :-
>>>>
>>>> B HDD from server 1
>>>> B HDD from server 2
>>>> B HDD from server 3
>>>>
>>>> But when i add more bricks as below :-
>>>>
>>>> -----------------------
>>>> [root@stor1 ~]# gluster volume info
>>>>
>>>> Volume Name: 3TB
>>>> Type: Distributed-Replicate
>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
>>>> Status: Stopped
>>>> Number of Bricks: 5 x 3 = 15
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: stor1:/bricks/b/vol1
>>>> Brick2: stor2:/bricks/b/vol1
>>>> Brick3: stor3:/bricks/b/vol1
>>>> Brick4: stor1:/bricks/c/vol1
>>>> Brick5: stor2:/bricks/c/vol1
>>>> Brick6: stor3:/bricks/c/vol1
>>>> Brick7: stor1:/bricks/d/vol1
>>>> Brick8: stor2:/bricks/d/vol1
>>>> Brick9: stor3:/bricks/d/vol1
>>>> Brick10: stor1:/bricks/e/vol1
>>>> Brick11: stor2:/bricks/e/vol1
>>>> Brick12: stor3:/bricks/e/vol1
>>>> Brick13: stor1:/bricks/f/vol1
>>>> Brick14: stor2:/bricks/f/vol1
>>>> Brick15: stor3:/bricks/f/vol1
>>>> Options Reconfigured:
>>>> nfs.disable: off
>>>> user.cifs: enable
>>>> auth.allow: *
>>>> performance.quick-read: off
>>>> performance.read-ahead: off
>>>> performance.io-cache: off
>>>> performance.stat-prefetch: off
>>>> cluster.eager-lock: enable
>>>> network.remote-dio: enable
>>>> cluster.quorum-type: auto
>>>> cluster.server-quorum-type: server
>>>> storage.owner-uid: 36
>>>> storage.owner-gid: 36
>>>> --------------------------------
>>>>
>>>> Brick added successfully without any error but after 1 min quorum failed
>>>> and gluster stop working...
>>>>
>>> Punit,
>>
>> And what do you mean by quorum failed ?
>> What is effect that you are seeing ?
>> Could you provide output of 'gluster volume status' as well ?
>>
>> -- Sathees
>>
>>
>>  What do log files say?
>>>
>>>> Thanks,
>>>> Punit
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users@xxxxxxxxxxx
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>
>>>>
>>
>

--
~Atin

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux