Re: Quorum For replica 3 storage

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Atin,

Yes..Now i reproduce the problem again...with only 3 bricks it's working fine...

Inline image 1
Inline image 2

Now when i added more 6 bricks in to the same volume..bricks added successfully,but after adding the bricks the type of the volume has been chaned from replicate to distributed replicate...

Inline image 3

Is there any body can help me in this...now it's more and more complex..

Thanks,
Punit

On Mon, Jul 20, 2015 at 5:12 PM, Ramesh Nachimuthu <rnachimu@xxxxxxxxxx> wrote:


On 07/20/2015 12:44 PM, Punit Dambiwal wrote:
Hi Atin,

Earlier i was adding the brick and create the volume through ovirt itself....this time i did it through commandline...i added all 15 bricks 5*3 through command line....

It seems before optimize for virt storage it all ok...but after that one quorum start and it failed the whole...


"Optimize for Virt" action in oVirt sets the following volume options.

group = > virt
storage.owner-uid => 36
storage.owner-gid => 36
server.allow-insecure => on
network.ping-timeout => 10

Regards,
Ramesh

On Mon, Jul 20, 2015 at 3:07 PM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:
In one of your earlier email you mentioned that after adding a brick
volume status stopped working. Can you point me to the glusterd log for
that transaction?

~Atin

On 07/20/2015 12:11 PM, Punit Dambiwal wrote:
> Hi Atin,
>
> Please find the below details :-
>
> [image: Inline image 1]
>
> [image: Inline image 2]
>
> Now when i set the optimize for the virt storage under ovirt and restart
> glusterd service on any node...it start failing the quorum..
>
> [image: Inline image 3]
>
> [image: Inline image 4]
>
> Thanks,
> Punit
>
> On Mon, Jul 20, 2015 at 10:44 AM, Punit Dambiwal <hypunit@xxxxxxxxx> wrote:
>
>> HI Atin,
>>
>> Apologies for the delay response...
>>
>> 1. When you added the brick was the command successful?
>>>> Yes..it was successful..
>> 2. If volume status is failing what's output its throwing in the console
>> and how about the glusterd log?
>>>> I will reproduce the issue again and update you..
>>
>> On Mon, Jul 13, 2015 at 11:46 AM, Atin Mukherjee <amukherj@xxxxxxxxxx>
>> wrote:
>>
>>>
>>>
>>> On 07/13/2015 05:19 AM, Punit Dambiwal wrote:
>>>> Hi Sathees,
>>>>
>>>> With 3 bricks i can get the gluster volume status....but after added
>>> more
>>>> bricks....can not get gluster volume status....
>>> The information is still incomplete in respect to analyze the problem.
>>> Further questions:
>>>
>>> 1. When you added the brick was the command successful?
>>> 2. If volume status is failing what's output its throwing in the console
>>> and how about the glusterd log?
>>>
>>> ~Atin
>>>>
>>>> On Sun, Jul 12, 2015 at 11:09 AM, SATHEESARAN <sasundar@xxxxxxxxxx>
>>> wrote:
>>>>
>>>>> On 07/11/2015 02:46 PM, Atin Mukherjee wrote:
>>>>>
>>>>>>
>>>>>> On 07/10/2015 03:03 PM, Punit Dambiwal wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> I have deployed one replica 3 storage...but i am facing some issue
>>> with
>>>>>>> quorum...
>>>>>>>
>>>>>>> Let me elaborate more :-
>>>>>>>
>>>>>>> 1. I have 3 node machines and every machine has 5 HDD(Bricks)...No
>>>>>>> RAID...Just JBOD...
>>>>>>> 2. Gluster working fine when just add 3 HDD as below :-
>>>>>>>
>>>>>>> B HDD from server 1
>>>>>>> B HDD from server 2
>>>>>>> B HDD from server 3
>>>>>>>
>>>>>>> But when i add more bricks as below :-
>>>>>>>
>>>>>>> -----------------------
>>>>>>> [root@stor1 ~]# gluster volume info
>>>>>>>
>>>>>>> Volume Name: 3TB
>>>>>>> Type: Distributed-Replicate
>>>>>>> Volume ID: 5be9165c-3402-4083-b3db-b782da2fb8d8
>>>>>>> Status: Stopped
>>>>>>> Number of Bricks: 5 x 3 = 15
>>>>>>> Transport-type: tcp
>>>>>>> Bricks:
>>>>>>> Brick1: stor1:/bricks/b/vol1
>>>>>>> Brick2: stor2:/bricks/b/vol1
>>>>>>> Brick3: stor3:/bricks/b/vol1
>>>>>>> Brick4: stor1:/bricks/c/vol1
>>>>>>> Brick5: stor2:/bricks/c/vol1
>>>>>>> Brick6: stor3:/bricks/c/vol1
>>>>>>> Brick7: stor1:/bricks/d/vol1
>>>>>>> Brick8: stor2:/bricks/d/vol1
>>>>>>> Brick9: stor3:/bricks/d/vol1
>>>>>>> Brick10: stor1:/bricks/e/vol1
>>>>>>> Brick11: stor2:/bricks/e/vol1
>>>>>>> Brick12: stor3:/bricks/e/vol1
>>>>>>> Brick13: stor1:/bricks/f/vol1
>>>>>>> Brick14: stor2:/bricks/f/vol1
>>>>>>> Brick15: stor3:/bricks/f/vol1
>>>>>>> Options Reconfigured:
>>>>>>> nfs.disable: off
>>>>>>> user.cifs: enable
>>>>>>> auth.allow: *
>>>>>>> performance.quick-read: off
>>>>>>> performance.read-ahead: off
>>>>>>> performance.io-cache: off
>>>>>>> performance.stat-prefetch: off
>>>>>>> cluster.eager-lock: enable
>>>>>>> network.remote-dio: enable
>>>>>>> cluster.quorum-type: auto
>>>>>>> cluster.server-quorum-type: server
>>>>>>> storage.owner-uid: 36
>>>>>>> storage.owner-gid: 36
>>>>>>> --------------------------------
>>>>>>>
>>>>>>> Brick added successfully without any error but after 1 min quorum
>>> failed
>>>>>>> and gluster stop working...
>>>>>>>
>>>>>> Punit,
>>>>>
>>>>> And what do you mean by quorum failed ?
>>>>> What is effect that you are seeing ?
>>>>> Could you provide output of 'gluster volume status' as well ?
>>>>>
>>>>> -- Sathees
>>>>>
>>>>>
>>>>>  What do log files say?
>>>>>>
>>>>>>> Thanks,
>>>>>>> Punit
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Gluster-users mailing list
>>>>>>> Gluster-users@xxxxxxxxxxx
>>>>>>> http://www.gluster.org/mailman/listinfo/gluster-users
>>>>>>>
>>>>>>>
>>>>>
>>>>
>>>
>>> --
>>> ~Atin
>>>
>>
>>
>

--
~Atin



_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users


_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://www.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux