Yes Artem! That's it!
I used the following commands and everything works as expected with 3 nodes:
gluster volume create VMS proxmox01:/DATA/vms
gluster vol start VMS
gluster vol status VMS
gluster peer probe proxmox02
gluster volume add-brick VMS replica 2 proxmox02:/DATA/vms
gluster vol status VMS
gluster vol info VMS
gluster peer probe proxmox03
gluster volume add-brick VMS replica 3 proxmox03:/DATA/vms
gluster vol set VMS cluster.heal-timeout 60
gluster volume heal VMS enable
gluster vol set VMS cluster.quorum-reads false
gluster vol set VMS cluster.quorum-count 1
gluster vol start VMS
gluster vol status VMS
gluster peer probe proxmox02
gluster volume add-brick VMS replica 2 proxmox02:/DATA/vms
gluster vol status VMS
gluster vol info VMS
gluster peer probe proxmox03
gluster volume add-brick VMS replica 3 proxmox03:/DATA/vms
gluster vol set VMS cluster.heal-timeout 60
gluster volume heal VMS enable
gluster vol set VMS cluster.quorum-reads false
gluster vol set VMS cluster.quorum-count 1
Thanks for you replay
Cheers
---
Gilberto Nunes Ferreira
Em sex., 17 de jul. de 2020 às 16:56, Artem Russakovskii <archon810@xxxxxxxxx> escreveu:
I had the same requirements (except with 4 servers and no arbiter), and this was the solution:gluster v set VMS cluster.quorum-count 1
gluster v set VMS cluster.quorum-type fixed
On Fri, Jul 17, 2020 at 6:50 AM Gilberto Nunes <gilberto.nunes32@xxxxxxxxx> wrote:How there________I have 3 servers with gluster 7 installed and setting up with replica 3 and arbiter 1.Here's the commands I used:- First create a simple volume with one server:gluster volume create VMS proxmox01:/DATA/vms- Then add the second one
gluster peer probe proxmox02
gluster volume add-brick VMS replica 2 proxmox02:/DATA/vms
- And finally and the third:gluster peer probe proxmox03gluster volume add-brick VMS replica 3 arbiter 1 proxmox03:/DATA/vmsBut then I decide to test the environment and bring proxmox02 and proxmox03 down and get Transport endpoint is not connected after a few seconds.Is there a way to keep one server up if 2 goes down?gluster vol info
Volume Name: VMS
Type: Replicate
Volume ID: 64735da4-8671-4c5e-b832-d15f5c03e9f0
Status: Started
Snapshot Count: 0
Number of Bricks: 1 x (2 + 1) = 3
Transport-type: tcp
Bricks:
Brick1: proxmox01:/DATA/vms
Brick2: proxmox02:/DATA/vms
Brick3: proxmox03:/DATA/vms (arbiter)
Options Reconfigured:
nfs.disable: on
storage.fips-mode-rchecksum: on
transport.address-family: inet
performance.client-io-threads: off
cluster.self-heal-daemon: enable
cluster.quorum-reads: false
cluster.quorum-count: 1
gluster vol status
Status of volume: VMS
Gluster process TCP Port RDMA Port Online Pid
------------------------------------------------------------------------------
Brick proxmox01:/DATA/vms 49152 0 Y 1526
Self-heal Daemon on localhost N/A N/A Y 1537
Task Status of Volume VMS
------------------------------------------------------------------------------
There are no active volume tasks
Thanks a lot---Gilberto Nunes Ferreira(47) 3025-5907(47) 99676-7530 - Whatsapp / TelegramSkype: gilberto.nunes36
Community Meeting Calendar:
Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users
________ Community Meeting Calendar: Schedule - Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC Bridge: https://bluejeans.com/441850968 Gluster-users mailing list Gluster-users@xxxxxxxxxxx https://lists.gluster.org/mailman/listinfo/gluster-users