Re: Add an arbiter when have multiple bricks at    same server.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Ok.
I have a 3rd host with Debian 12 installed and Gluster v11. The name of the host is arbiter!

I already add this host into the pool:
arbiter:~# gluster pool list
UUID                                    Hostname                State
0cbbfc27-3876-400a-ac1d-2d73e72a4bfd    gluster1.home.local     Connected  
99ed1f1e-7169-4da8-b630-a712a5b71ccd    gluster2                Connected  
4718ead7-aebd-4b8b-a401-f9e8b0acfeb1    localhost               Connected

But when I do this:
pve01:~# gluster volume add-brick VMS replica 2 arbiter 1 arbiter:/arbiter1 arbiter:/arbiter2 arbiter:/arbiter3
I got this error:

For arbiter configuration, replica count must be 3 and arbiter count must be 1. The 3rd brick of the replica will be the arbiter

Usage:
volume add-brick <VOLNAME> [<replica> <COUNT> [arbiter <COUNT>]] <NEW-BRICK> ... [force]

gluster vol info
pve01:~# gluster vol info
 
Volume Name: VMS
Type: Distributed-Replicate
Volume ID: e1a4f787-3f62-441e-a7ce-c0ae6b111ebf
Status: Started
Snapshot Count: 0
Number of Bricks: 3 x 2 = 6
Transport-type: tcp
Bricks:
Brick1: gluster1:/disco2TB-0/vms
Brick2: gluster2:/disco2TB-0/vms
Brick3: gluster1:/disco1TB-0/vms
Brick4: gluster2:/disco1TB-0/vms
Brick5: gluster1:/disco1TB-1/vms
Brick6: gluster2:/disco1TB-1/vms
Options Reconfigured:
performance.client-io-threads: off
transport.address-family: inet
storage.fips-mode-rchecksum: on
cluster.granular-entry-heal: on
cluster.data-self-heal: off
cluster.metadata-self-heal: off
cluster.entry-self-heal: off
cluster.self-heal-daemon: off

What am I doing wrong?




---


Gilberto Nunes Ferreira
(47) 99676-7530 - Whatsapp / Telegram






Em qua., 6 de nov. de 2024 às 11:32, Strahil Nikolov <hunter86_bg@xxxxxxxxx> escreveu:
Right now you have 3 "sets" of replica 2 on 2 hosts.
In your case you don't need so much space for arbiters (10-15GB with 95 maxpct is enough for each "set") and you need a 3rd system or when the node that holds the data brick + arbiter brick fails (2 node scenario) - that "set" will be unavailable.

If you do have a 3rd host, I think the command would be:
gluster volume add-brick VOLUME replica 2 arbiter 1 server3:/first/set/arbiter server3:/second/set/arbiter server3:/last/set/arbiter


Best Regards,
Strahil Nikolov

Best Regards,
Strahil Nikolov

On Tue, Nov 5, 2024 at 21:17, Gilberto Ferreira
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux