Ok...Thanks a lot Strahil
This gluster volume set VMS cluster.favorite-child-policy size do the trick to me here!
Cheers
---
Gilberto Nunes Ferreira
(47) 3025-5907
(47) 99676-7530 - Whatsapp / Telegram
Skype: gilberto.nunes36
Em qua., 5 de ago. de 2020 às 18:15, Strahil Nikolov <hunter86_bg@xxxxxxxxx> escreveu:
This could happen if you have pending heals. Did you reboot that node recently ?
Did you set automatic unsplit-brain ?
Check for pending heals and files in splitbrain.
If not, you can check https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/ (look at point 5).
Best Regards,
Strahil Nikolov
На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes <gilberto.nunes32@xxxxxxxxx> написа:
>I'm in trouble here.
>When I shutdown the pve01 server, the shared folder over glusterfs is
>EMPTY!
>It's supposed to be a qcow2 file inside it.
>The content is show right, just after I power on pve01 backup...
>
>Some advice?
>
>
>Thanks
>
>---
>Gilberto Nunes Ferreira
>
>(47) 3025-5907
>(47) 99676-7530 - Whatsapp / Telegram
>
>Skype: gilberto.nunes36
>
>
>
>
>
>Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes <
>gilberto.nunes32@xxxxxxxxx> escreveu:
>
>> Well...
>> I do the follow:
>>
>> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1
>> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv
>> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force
>>
>> And now I have:
>> gluster vol info
>>
>> Volume Name: VMS
>> Type: Distributed-Replicate
>> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 2 x (2 + 1) = 6
>> Transport-type: tcp
>> Bricks:
>> Brick1: pve01:/DATA/brick1
>> Brick2: pve02:/DATA/brick1.5
>> Brick3: pve01:/DATA/arbiter1.5 (arbiter)
>> Brick4: pve02:/DATA/brick2
>> Brick5: pve01:/DATA/brick2.5
>> Brick6: pve02:/DATA/arbiter2.5 (arbiter)
>> Options Reconfigured:
>> cluster.quorum-count: 1
>> cluster.quorum-reads: false
>> cluster.self-heal-daemon: enable
>> cluster.heal-timeout: 10
>> storage.fips-mode-rchecksum: on
>> transport.address-family: inet
>> nfs.disable: on
>> performance.client-io-threads: off
>>
>> This values I have put it myself, in order to see if could improve
>the
>> time to make the volume available, when pve01 goes down with ifupdown
>> cluster.quorum-count: 1
>> cluster.quorum-reads: false
>> cluster.self-heal-daemon: enable
>> cluster.heal-timeout: 10
>>
>> Nevertheless, it took more than 1 minutes to the volume VMS available
>in
>> the other host (pve02).
>> Is there any trick to reduce this time ?
>>
>> Thanks
>>
>> ---
>> Gilberto Nunes Ferreira
>>
>>
>>
>>
>>
>>
>> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes <
>> gilberto.nunes32@xxxxxxxxx> escreveu:
>>
>>> hum I see... like this:
>>> [image: image.png]
>>> ---
>>> Gilberto Nunes Ferreira
>>>
>>> (47) 3025-5907
>>> (47) 99676-7530 - Whatsapp / Telegram
>>>
>>> Skype: gilberto.nunes36
>>>
>>>
>>>
>>>
>>>
>>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation <
>>> bob@xxxxxxxxxxxxxxx> escreveu:
>>>
>>>> check the example of the chained configuration on this page:
>>>>
>>>>
>>>>
>https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes
>>>>
>>>> and apply it to two servers...
>>>>
>>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote:
>>>> > Hi Bob!
>>>> >
>>>> > Could you, please, send me more detail about this configuration?
>>>> > I will appreciate that!
>>>> >
>>>> > Thank you
>>>> > ---
>>>> > Gilberto Nunes Ferreira
>>>> >
>>>> > (47) 3025-5907
>>>> > **
>>>> > (47) 99676-7530 - Whatsapp / Telegram
>>>> >
>>>> > Skype: gilberto.nunes36
>>>> >
>>>> >
>>>> >
>>>> >
>>>> >
>>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation
>>>> > <bob@xxxxxxxxxxxxxxx <mailto:bob@xxxxxxxxxxxxxxx>> escreveu:
>>>> >
>>>> > Hi Gilberto,
>>>> >
>>>> > My understanding is there can only be one arbiter per
>replicated
>>>> > set. I
>>>> > don't have a lot of practice with gluster, so this could be
>bad
>>>> advice,
>>>> > but the way I dealt with it on my two servers was to use 6
>bricks
>>>> as
>>>> > distributed-replicated (this is also relatively easy to
>migrate to
>>>> 3
>>>> > servers if that happens for you in the future):
>>>> >
>>>> > Server1 Server2
>>>> > brick1 brick1.5
>>>> > arbiter1.5 brick2
>>>> > brick2.5 arbiter2.5
>>>> >
>>>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote:
>>>> > > Hi there.
>>>> > > I have two physical servers deployed as replica 2 and,
>>>> obviously,
>>>> > I got
>>>> > > a split-brain.
>>>> > > So I am thinking in use two virtual machines,each one in
>>>> physical
>>>> > > servers....
>>>> > > Then this two VMS act as a artiber of gluster set....
>>>> > >
>>>> > > Is this doable?
>>>> > >
>>>> > > Thanks
>>>> > >
>>>> > > ________
>>>> > >
>>>> > >
>>>> > >
>>>> > > Community Meeting Calendar:
>>>> > >
>>>> > > Schedule -
>>>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>>> > > Bridge: https://bluejeans.com/441850968
>>>> > >
>>>> > > Gluster-users mailing list
>>>> > > Gluster-users@xxxxxxxxxxx
><mailto:Gluster-users@xxxxxxxxxxx>
>>>> > > https://lists.gluster.org/mailman/listinfo/gluster-users
>>>> > >
>>>> > ________
>>>> >
>>>> >
>>>> >
>>>> > Community Meeting Calendar:
>>>> >
>>>> > Schedule -
>>>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
>>>> > Bridge: https://bluejeans.com/441850968
>>>> >
>>>> > Gluster-users mailing list
>>>> > Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx>
>>>> > https://lists.gluster.org/mailman/listinfo/gluster-users
>>>> >
>>>>
>>>
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users