This could happen if you have pending heals. Did you reboot that node recently ? Did you set automatic unsplit-brain ? Check for pending heals and files in splitbrain. If not, you can check https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/ (look at point 5). Best Regards, Strahil Nikolov На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes <gilberto.nunes32@xxxxxxxxx> написа: >I'm in trouble here. >When I shutdown the pve01 server, the shared folder over glusterfs is >EMPTY! >It's supposed to be a qcow2 file inside it. >The content is show right, just after I power on pve01 backup... > >Some advice? > > >Thanks > >--- >Gilberto Nunes Ferreira > >(47) 3025-5907 >(47) 99676-7530 - Whatsapp / Telegram > >Skype: gilberto.nunes36 > > > > > >Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes < >gilberto.nunes32@xxxxxxxxx> escreveu: > >> Well... >> I do the follow: >> >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force >> >> And now I have: >> gluster vol info >> >> Volume Name: VMS >> Type: Distributed-Replicate >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd >> Status: Started >> Snapshot Count: 0 >> Number of Bricks: 2 x (2 + 1) = 6 >> Transport-type: tcp >> Bricks: >> Brick1: pve01:/DATA/brick1 >> Brick2: pve02:/DATA/brick1.5 >> Brick3: pve01:/DATA/arbiter1.5 (arbiter) >> Brick4: pve02:/DATA/brick2 >> Brick5: pve01:/DATA/brick2.5 >> Brick6: pve02:/DATA/arbiter2.5 (arbiter) >> Options Reconfigured: >> cluster.quorum-count: 1 >> cluster.quorum-reads: false >> cluster.self-heal-daemon: enable >> cluster.heal-timeout: 10 >> storage.fips-mode-rchecksum: on >> transport.address-family: inet >> nfs.disable: on >> performance.client-io-threads: off >> >> This values I have put it myself, in order to see if could improve >the >> time to make the volume available, when pve01 goes down with ifupdown >> cluster.quorum-count: 1 >> cluster.quorum-reads: false >> cluster.self-heal-daemon: enable >> cluster.heal-timeout: 10 >> >> Nevertheless, it took more than 1 minutes to the volume VMS available >in >> the other host (pve02). >> Is there any trick to reduce this time ? >> >> Thanks >> >> --- >> Gilberto Nunes Ferreira >> >> >> >> >> >> >> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes < >> gilberto.nunes32@xxxxxxxxx> escreveu: >> >>> hum I see... like this: >>> [image: image.png] >>> --- >>> Gilberto Nunes Ferreira >>> >>> (47) 3025-5907 >>> (47) 99676-7530 - Whatsapp / Telegram >>> >>> Skype: gilberto.nunes36 >>> >>> >>> >>> >>> >>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation < >>> bob@xxxxxxxxxxxxxxx> escreveu: >>> >>>> check the example of the chained configuration on this page: >>>> >>>> >>>> >https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes >>>> >>>> and apply it to two servers... >>>> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: >>>> > Hi Bob! >>>> > >>>> > Could you, please, send me more detail about this configuration? >>>> > I will appreciate that! >>>> > >>>> > Thank you >>>> > --- >>>> > Gilberto Nunes Ferreira >>>> > >>>> > (47) 3025-5907 >>>> > ** >>>> > (47) 99676-7530 - Whatsapp / Telegram >>>> > >>>> > Skype: gilberto.nunes36 >>>> > >>>> > >>>> > >>>> > >>>> > >>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation >>>> > <bob@xxxxxxxxxxxxxxx <mailto:bob@xxxxxxxxxxxxxxx>> escreveu: >>>> > >>>> > Hi Gilberto, >>>> > >>>> > My understanding is there can only be one arbiter per >replicated >>>> > set. I >>>> > don't have a lot of practice with gluster, so this could be >bad >>>> advice, >>>> > but the way I dealt with it on my two servers was to use 6 >bricks >>>> as >>>> > distributed-replicated (this is also relatively easy to >migrate to >>>> 3 >>>> > servers if that happens for you in the future): >>>> > >>>> > Server1 Server2 >>>> > brick1 brick1.5 >>>> > arbiter1.5 brick2 >>>> > brick2.5 arbiter2.5 >>>> > >>>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: >>>> > > Hi there. >>>> > > I have two physical servers deployed as replica 2 and, >>>> obviously, >>>> > I got >>>> > > a split-brain. >>>> > > So I am thinking in use two virtual machines,each one in >>>> physical >>>> > > servers.... >>>> > > Then this two VMS act as a artiber of gluster set.... >>>> > > >>>> > > Is this doable? >>>> > > >>>> > > Thanks >>>> > > >>>> > > ________ >>>> > > >>>> > > >>>> > > >>>> > > Community Meeting Calendar: >>>> > > >>>> > > Schedule - >>>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>>> > > Bridge: https://bluejeans.com/441850968 >>>> > > >>>> > > Gluster-users mailing list >>>> > > Gluster-users@xxxxxxxxxxx ><mailto:Gluster-users@xxxxxxxxxxx> >>>> > > https://lists.gluster.org/mailman/listinfo/gluster-users >>>> > > >>>> > ________ >>>> > >>>> > >>>> > >>>> > Community Meeting Calendar: >>>> > >>>> > Schedule - >>>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >>>> > Bridge: https://bluejeans.com/441850968 >>>> > >>>> > Gluster-users mailing list >>>> > Gluster-users@xxxxxxxxxxx <mailto:Gluster-users@xxxxxxxxxxx> >>>> > https://lists.gluster.org/mailman/listinfo/gluster-users >>>> > >>>> >>> ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users