Re: Two VMS as arbiter...

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



If I understood you correctly, you are looking for this:
Option: network.ping-timeout                    Default Value: 42                               Description: Time duration for which the client waits to check if the server is responsive.

Best Regards,
Strahil Nikolov


На 5 август 2020 г. 17:07:10 GMT+03:00, Gilberto Nunes <gilberto.nunes32@xxxxxxxxx> написа:
>Well...
>I do the follow:
>
>gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1
>pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv
>e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
>
>And now I have:
>gluster vol info
>
>Volume Name: VMS
>Type: Distributed-Replicate
>Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
>Status: Started
>Snapshot Count: 0
>Number of Bricks: 2 x (2 + 1) = 6
>Transport-type: tcp
>Bricks:
>Brick1: pve01:/DATA/brick1
>Brick2: pve02:/DATA/brick1.5
>Brick3: pve01:/DATA/arbiter1.5 (arbiter)
>Brick4: pve02:/DATA/brick2
>Brick5: pve01:/DATA/brick2.5
>Brick6: pve02:/DATA/arbiter2.5 (arbiter)
>Options Reconfigured:
>cluster.quorum-count: 1
>cluster.quorum-reads: false
>cluster.self-heal-daemon: enable
>cluster.heal-timeout: 10
>storage.fips-mode-rchecksum: on
>transport.address-family: inet
>nfs.disable: on
>performance.client-io-threads: off
>
>This values I have put it myself, in order to see if could improve the
>time
>to make the volume available, when pve01 goes down with ifupdown
>cluster.quorum-count: 1
>cluster.quorum-reads: false
>cluster.self-heal-daemon: enable
>cluster.heal-timeout: 10
>
>Nevertheless, it took more than 1 minutes to the volume VMS available
>in
>the other host (pve02).
>Is there any trick to reduce this time ?
>
>Thanks
>
>---
>Gilberto Nunes Ferreira
>
>
>
>
>
>
>Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes <
>gilberto.nunes32@xxxxxxxxx> escreveu:
>
>> hum I see... like this:
>> [image: image.png]
>> ---
>> Gilberto Nunes Ferreira
>>
>> (47) 3025-5907
>> (47) 99676-7530 - Whatsapp / Telegram
>>
>> Skype: gilberto.nunes36
>>
>>
>>
>>
>>
>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation <
>> bob@xxxxxxxxxxxxxxx> escreveu:
>>
>>> check the example of the chained configuration on this page:
>>>
>>>
>>>
>https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
>>>
>>> and apply it to two servers...
>>>
>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote:
>>> > Hi Bob!
>>> >
>>> > Could you, please, send me more detail about this configuration?
>>> > I will appreciate that!
>>> >
>>> > Thank you
>>> > ---
>>> > Gilberto Nunes Ferreira
>>> >
>>> > (47) 3025-5907
>>> > **
>>> > (47) 99676-7530 - Whatsapp / Telegram
>>> >
>>> > Skype: gilberto.nunes36
>>> >
>>> >
>>> >
>>> >
>>> >
>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation
>>> > <bob@xxxxxxxxxxxxxxx <mailto:bob@xxxxxxxxxxxxxxx>> escreveu:
>>> >
>>> >     Hi Gilberto,
>>> >
>>> >     My understanding is there can only be one arbiter per
>replicated
>>> >     set.  I
>>> >     don't have a lot of practice with gluster, so this could be
>bad
>>> advice,
>>> >     but the way I dealt with it on my two servers was to use 6
>bricks as
>>> >     distributed-replicated (this is also relatively easy to
>migrate to 3
>>> >     servers if that happens for you in the future):
>>> >
>>> >     Server1     Server2
>>> >     brick1      brick1.5
>>> >     arbiter1.5  brick2
>>> >     brick2.5    arbiter2.5
>>> >
>>> >     On 2020-08-04 7:00 p.m., Gilberto Nunes wrote:
>>> >      > Hi there.
>>> >      > I have two physical servers deployed as replica 2 and,
>obviously,
>>> >     I got
>>> >      > a split-brain.
>>> >      > So I am thinking in use two virtual machines,each one in
>physical
>>> >      > servers....
>>> >      > Then this two VMS act as a artiber of gluster set....
>>> >      >
>>> >      > Is this doable?
>>> >      >
>>> >      > Thanks
>>> >      >
>>> >      > ________
>>> >      >
>>> >      >
>>> >      >
>>> >      > Community Meeting Calendar:
>>> >      >
>>> >      > Schedule -
>>> >      > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>> >      > Bridge: https://bluejeans.com/441850968
>>> >      >
>>> >      > Gluster-users mailing list
>>> >      > Gluster-users@xxxxxxxxxxx
><mailto:Gluster-users@xxxxxxxxxxx>
>>> >      > https://lists.gluster.org/mailman/listinfo/gluster-users
>>> >      >
>>> >     ________
>>> >
>>> >
>>> >
>>> >     Community Meeting Calendar:
>>> >
>>> >     Schedule -
>>> >     Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>> >     Bridge: https://bluejeans.com/441850968
>>> >
>>> >     Gluster-users mailing list
>>> >     Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx>
>>> >     https://lists.gluster.org/mailman/listinfo/gluster-users
>>> >
>>>
>>
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users




[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux