As you mentioned qcow2 files, check the virt group (/var/lib/glusterfs/group or something like that). It has optimal setttins for VMs and is used by oVirt. WARNING: If you decide to enable the group, which will also enable sharding, NEVER EVER DISABLE SHARDING -> ONCE ENABLED STAYS ENABLED !!! Sharding helps reduce loocking during replica heals. WARNING2: As virt group uses sharding (fixes the size of file into shard size), you should consider cluster.favorite-child-policy with value ctime/mtime. Best Regards, Strahil Nikolov На 6 август 2020 г. 1:56:58 GMT+03:00, Gilberto Nunes <gilberto.nunes32@xxxxxxxxx> написа: >Ok...Thanks a lot Strahil > >This gluster volume set VMS cluster.favorite-child-policy size do the >trick >to me here! > >Cheers >--- >Gilberto Nunes Ferreira > >(47) 3025-5907 >(47) 99676-7530 - Whatsapp / Telegram > >Skype: gilberto.nunes36 > > > > > >Em qua., 5 de ago. de 2020 às 18:15, Strahil Nikolov ><hunter86_bg@xxxxxxxxx> >escreveu: > >> This could happen if you have pending heals. Did you reboot that node >> recently ? >> Did you set automatic unsplit-brain ? >> >> Check for pending heals and files in splitbrain. >> >> If not, you can check >> >https://docs.gluster.org/en/latest/Troubleshooting/resolving-splitbrain/ >> (look at point 5). >> >> Best Regards, >> Strahil Nikolov >> >> На 5 август 2020 г. 23:41:57 GMT+03:00, Gilberto Nunes < >> gilberto.nunes32@xxxxxxxxx> написа: >> >I'm in trouble here. >> >When I shutdown the pve01 server, the shared folder over glusterfs >is >> >EMPTY! >> >It's supposed to be a qcow2 file inside it. >> >The content is show right, just after I power on pve01 backup... >> > >> >Some advice? >> > >> > >> >Thanks >> > >> >--- >> >Gilberto Nunes Ferreira >> > >> >(47) 3025-5907 >> >(47) 99676-7530 - Whatsapp / Telegram >> > >> >Skype: gilberto.nunes36 >> > >> > >> > >> > >> > >> >Em qua., 5 de ago. de 2020 às 11:07, Gilberto Nunes < >> >gilberto.nunes32@xxxxxxxxx> escreveu: >> > >> >> Well... >> >> I do the follow: >> >> >> >> gluster vol create VMS replica 3 arbiter 1 pve01:/DATA/brick1 >> >> pve02:/DATA/brick1.5 pve01:/DATA/arbiter1.5 pve02:/DATA/brick2 pv >> >> e01:/DATA/brick2.5 pve02:/DATA/arbiter2.5 force >> >> >> >> And now I have: >> >> gluster vol info >> >> >> >> Volume Name: VMS >> >> Type: Distributed-Replicate >> >> Volume ID: 1bd712f5-ccb9-4322-8275-abe363d1ffdd >> >> Status: Started >> >> Snapshot Count: 0 >> >> Number of Bricks: 2 x (2 + 1) = 6 >> >> Transport-type: tcp >> >> Bricks: >> >> Brick1: pve01:/DATA/brick1 >> >> Brick2: pve02:/DATA/brick1.5 >> >> Brick3: pve01:/DATA/arbiter1.5 (arbiter) >> >> Brick4: pve02:/DATA/brick2 >> >> Brick5: pve01:/DATA/brick2.5 >> >> Brick6: pve02:/DATA/arbiter2.5 (arbiter) >> >> Options Reconfigured: >> >> cluster.quorum-count: 1 >> >> cluster.quorum-reads: false >> >> cluster.self-heal-daemon: enable >> >> cluster.heal-timeout: 10 >> >> storage.fips-mode-rchecksum: on >> >> transport.address-family: inet >> >> nfs.disable: on >> >> performance.client-io-threads: off >> >> >> >> This values I have put it myself, in order to see if could improve >> >the >> >> time to make the volume available, when pve01 goes down with >ifupdown >> >> cluster.quorum-count: 1 >> >> cluster.quorum-reads: false >> >> cluster.self-heal-daemon: enable >> >> cluster.heal-timeout: 10 >> >> >> >> Nevertheless, it took more than 1 minutes to the volume VMS >available >> >in >> >> the other host (pve02). >> >> Is there any trick to reduce this time ? >> >> >> >> Thanks >> >> >> >> --- >> >> Gilberto Nunes Ferreira >> >> >> >> >> >> >> >> >> >> >> >> >> >> Em qua., 5 de ago. de 2020 às 08:57, Gilberto Nunes < >> >> gilberto.nunes32@xxxxxxxxx> escreveu: >> >> >> >>> hum I see... like this: >> >>> [image: image.png] >> >>> --- >> >>> Gilberto Nunes Ferreira >> >>> >> >>> (47) 3025-5907 >> >>> (47) 99676-7530 - Whatsapp / Telegram >> >>> >> >>> Skype: gilberto.nunes36 >> >>> >> >>> >> >>> >> >>> >> >>> >> >>> Em qua., 5 de ago. de 2020 às 02:14, Computerisms Corporation < >> >>> bob@xxxxxxxxxxxxxxx> escreveu: >> >>> >> >>>> check the example of the chained configuration on this page: >> >>>> >> >>>> >> >>>> >> > >> >https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.3/html/administration_guide/creating_arbitrated_replicated_volumes >> >>>> >> >>>> and apply it to two servers... >> >>>> >> >>>> On 2020-08-04 8:25 p.m., Gilberto Nunes wrote: >> >>>> > Hi Bob! >> >>>> > >> >>>> > Could you, please, send me more detail about this >configuration? >> >>>> > I will appreciate that! >> >>>> > >> >>>> > Thank you >> >>>> > --- >> >>>> > Gilberto Nunes Ferreira >> >>>> > >> >>>> > (47) 3025-5907 >> >>>> > ** >> >>>> > (47) 99676-7530 - Whatsapp / Telegram >> >>>> > >> >>>> > Skype: gilberto.nunes36 >> >>>> > >> >>>> > >> >>>> > >> >>>> > >> >>>> > >> >>>> > Em ter., 4 de ago. de 2020 às 23:47, Computerisms Corporation >> >>>> > <bob@xxxxxxxxxxxxxxx <mailto:bob@xxxxxxxxxxxxxxx>> escreveu: >> >>>> > >> >>>> > Hi Gilberto, >> >>>> > >> >>>> > My understanding is there can only be one arbiter per >> >replicated >> >>>> > set. I >> >>>> > don't have a lot of practice with gluster, so this could >be >> >bad >> >>>> advice, >> >>>> > but the way I dealt with it on my two servers was to use 6 >> >bricks >> >>>> as >> >>>> > distributed-replicated (this is also relatively easy to >> >migrate to >> >>>> 3 >> >>>> > servers if that happens for you in the future): >> >>>> > >> >>>> > Server1 Server2 >> >>>> > brick1 brick1.5 >> >>>> > arbiter1.5 brick2 >> >>>> > brick2.5 arbiter2.5 >> >>>> > >> >>>> > On 2020-08-04 7:00 p.m., Gilberto Nunes wrote: >> >>>> > > Hi there. >> >>>> > > I have two physical servers deployed as replica 2 and, >> >>>> obviously, >> >>>> > I got >> >>>> > > a split-brain. >> >>>> > > So I am thinking in use two virtual machines,each one >in >> >>>> physical >> >>>> > > servers.... >> >>>> > > Then this two VMS act as a artiber of gluster set.... >> >>>> > > >> >>>> > > Is this doable? >> >>>> > > >> >>>> > > Thanks >> >>>> > > >> >>>> > > ________ >> >>>> > > >> >>>> > > >> >>>> > > >> >>>> > > Community Meeting Calendar: >> >>>> > > >> >>>> > > Schedule - >> >>>> > > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> >>>> > > Bridge: https://bluejeans.com/441850968 >> >>>> > > >> >>>> > > Gluster-users mailing list >> >>>> > > Gluster-users@xxxxxxxxxxx >> ><mailto:Gluster-users@xxxxxxxxxxx> >> >>>> > > >https://lists.gluster.org/mailman/listinfo/gluster-users >> >>>> > > >> >>>> > ________ >> >>>> > >> >>>> > >> >>>> > >> >>>> > Community Meeting Calendar: >> >>>> > >> >>>> > Schedule - >> >>>> > Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC >> >>>> > Bridge: https://bluejeans.com/441850968 >> >>>> > >> >>>> > Gluster-users mailing list >> >>>> > Gluster-users@xxxxxxxxxxx ><mailto:Gluster-users@xxxxxxxxxxx> >> >>>> > https://lists.gluster.org/mailman/listinfo/gluster-users >> >>>> > >> >>>> >> >>> >> ________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users