HI Atin,
Apologies for the delay response...
1. When you added the brick was the command successful?
>> Yes..it was successful..
2. If volume status is failing what's output its throwing in the console
and how about the glusterd log?
2. If volume status is failing what's output its throwing in the console
and how about the glusterd log?
>> I will reproduce the issue again and update you..
On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:
On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
> Hi Sathees,
>
> With 3 bricks i can get the gluster volume status....but after added more
> bricks....can not get gluster volume status....
The information is still incomplete in respect to analyze the problem.
Further questions:
1. When you added the brick was the command successful?
2. If volume status is failing what's output its throwing in the console
and how about the glusterd log?
~Atin
-->
> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar@xxxxxxxxxx> wrote:
>
>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
>>
>>>
>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
>>>
>>>> Hi,
>>>>
>>>> I have deployed one replica 3 storage...but i am facing some issue with
>>>> quorum...
>>>>
>>>> Let me elaborate more :-
>>>>
>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No
>>>> RAID...Just JBOD...
>>>> 2. Gluster working fine when just add 3 HDD as below :-
>>>>
>>>> B HDD from server 1
>>>> B HDD from server 2
>>>> B HDD from server 3
>>>>
>>>> But when i add more bricks as below :-
>>>>
>>>> -----------------------
>>>> [root@stor1 ~]# gluster volume info
>>>>
>>>> Volume Name: 3TB
>>>> Type: Distributed-Replicate
>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
>>>> Status: Stopped
>>>> Number of Bricks: 5 x 3 = 15
>>>> Transport-type: tcp
>>>> Bricks:
>>>> Brick1: stor1:/bricks/b/vol1
>>>> Brick2: stor2:/bricks/b/vol1
>>>> Brick3: stor3:/bricks/b/vol1
>>>> Brick4: stor1:/bricks/c/vol1
>>>> Brick5: stor2:/bricks/c/vol1
>>>> Brick6: stor3:/bricks/c/vol1
>>>> Brick7: stor1:/bricks/d/vol1
>>>> Brick8: stor2:/bricks/d/vol1
>>>> Brick9: stor3:/bricks/d/vol1
>>>> Brick10: stor1:/bricks/e/vol1
>>>> Brick11: stor2:/bricks/e/vol1
>>>> Brick12: stor3:/bricks/e/vol1
>>>> Brick13: stor1:/bricks/f/vol1
>>>> Brick14: stor2:/bricks/f/vol1
>>>> Brick15: stor3:/bricks/f/vol1
>>>> Options Reconfigured:
>>>> nfs.disable: off
>>>> user.cifs: enable
>>>> auth.allow: *
>>>> performance.quick-read: off
>>>> performance.read-ahead: off
>>>> performance.io-cache: off
>>>> performance.stat-prefetch: off
>>>> cluster.eager-lock: enable
>>>> network.remote-dio: enable
>>>> cluster.quorum-type: auto
>>>> cluster.server-quorum-type: server
>>>> storage.owner-uid: 36
>>>> storage.owner-gid: 36
>>>> --------------------------------
>>>>
>>>> Brick added successfully without any error but after 1 min quorum failed
>>>> and gluster stop working...
>>>>
>>> Punit,
>>
>> And what do you mean by quorum failed ?
>> What is effect that you are seeing ?
>> Could you provide output of 'gluster volume status' as well ?
>>
>> -- Sathees
>>
>>
>> What do log files say?
>>>
>>>> Thanks,
>>>> Punit
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users@xxxxxxxxxxx
>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>
>>>>
>>
>
~Atin
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users