Re: [ovirt-users] Re: Single instance scaleup.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all.

I acrtually did something like that myself.

I started out with a single node HC cluster. I then added another node (and plan to add a third). This is what I did:

1) Set up the new node. Make sure that you have all dependencies. (In my case I started with a Centos 8 machine, and installed vdms-gluster and gluster-ansible)
2) Configure the bricks. For this I just copied over hc_wizard_inventory.yml over from the first node, edited it to fit the second node, and just ran the gluster.infra role.
3) Expand the volume. In this case with the following command:
gluster volume add-brick engine replica 2 <host>:/gluster_bricks/engine/engine
4) now just add the host as a hypervisor using the management console.

I plan on adding a third node. Then I want to have full replica on the engine, and replica 2 + arbiter on the vmstore volume.

Expanding gluster volumes, migrating from distributed to replicated and even replacing bricks etc. is rather easy in Gluster once you know how it works. I have even replaced all the servers on a live gluster cluster, without service interruption…

Krist

On Jul 18, 2019, 09:58 +0200, Leo David <leoalex@xxxxxxxxx>, wrote:
Hi,
Looks like the only way arround would be to create a brand-new volume as replicated on other disks, and start moving the vms all around the place between volumes ?
Cheers,

Leo

On Mon, May 27, 2019 at 1:53 PM Leo David <leoalex@xxxxxxxxx> wrote:
Hi,
Any suggestions ?
Thank you very much !

Leo

On Sun, May 26, 2019 at 4:38 PM Strahil Nikolov <hunter86_bg@xxxxxxxxx> wrote:
Yeah,
it seems different from the docs.
I'm adding the gluster users list ,as they are more experienced into that.

@Gluster-users,

can you provide some hint how to add aditional replicas to the below volumes , so they become 'replica 2 arbiter 1' or 'replica 3' type volumes ?


Best Regards,
Strahil Nikolov

В неделя, 26 май 2019 г., 15:16:18 ч. Гринуич+3, Leo David <leoalex@xxxxxxxxx> написа:


Thank you Strahil,
The engine and ssd-samsung are distributed...
So these are the ones that I need to have replicated accross new nodes.
I am not very sure about the procedure to accomplish this.
Thanks,

Leo

On Sun, May 26, 2019, 13:04 Strahil <hunter86_bg@xxxxxxxxx> wrote:

Hi Leo,
As you do not have a distributed volume , you can easily switch to replica 2 arbiter 1 or replica 3 volumes.

You can use the following for adding the bricks:

https://access.redhat.com/documentation/en-US/Red_Hat_Storage/2.1/html/Administration_Guide/Expanding_Volumes.html

Best Regards,
Strahil Nikoliv

On May 26, 2019 10:54, Leo David <leoalex@xxxxxxxxx> wrote:
Hi Stahil,
Thank you so much for yout input !

 gluster volume info


Volume Name: engine
Type: Distribute
Volume ID: d7449fc2-cc35-4f80-a776-68e4a3dbd7e1
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: 192.168.80.191:/gluster_bricks/engine/engine
Options Reconfigured:
nfs.disable: on
transport.address-family: inet
storage.owner-uid: 36
storage.owner-gid: 36
features.shard: on
performance.low-prio-threads: 32
performance.strict-o-direct: off
network.remote-dio: off
network.ping-timeout: 30
user.cifs: off
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
cluster.eager-lock: enable
Volume Name: ssd-samsung
Type: Distribute
Volume ID: 76576cc6-220b-4651-952d-99846178a19e
Status: Started
Snapshot Count: 0
Number of Bricks: 1
Transport-type: tcp
Bricks:
Brick1: 192.168.80.191:/gluster_bricks/sdc/data
Options Reconfigured:
cluster.eager-lock: enable
performance.io-cache: off
performance.read-ahead: off
performance.quick-read: off
user.cifs: off
network.ping-timeout: 30
network.remote-dio: off
performance.strict-o-direct: on
performance.low-prio-threads: 32
features.shard: on
storage.owner-gid: 36
storage.owner-uid: 36
transport.address-family: inet
nfs.disable: on

The other two hosts will be 192.168.80.192/193  - this is gluster dedicated network over 10GB sfp+ switch.
- host 2 wil have identical harware configuration with host 1 ( each disk is actually a raid0 array )
- host 3 has:
   -  1 ssd for OS
   -  1 ssd - for adding to engine volume in a full replica 3
   -  2 ssd's in a raid 1 array to be added as arbiter for the data volume ( ssd-samsung )
So the plan is to have "engine"  scaled in a full replica 3,  and "ssd-samsung" scalled in a replica 3 arbitrated.




On Sun, May 26, 2019 at 10:34 AM Strahil <hunter86_bg@xxxxxxxxx> wrote:

Hi Leo,

Gluster is quite smart, but in order to provide any hints , can you provide output of 'gluster volume info <glustervol>'.
If you have 2 more systems , keep in mind that it is best to mirror the storage on the second replica (2 disks on 1 machine -> 2 disks on the new machine), while for the arbiter this is not neccessary.

What is your network and NICs ? Based on my experience , I can recommend at least 10 gbit/s  interfase(s).

Best Regards,
Strahil Nikolov

On May 26, 2019 07:52, Leo David <leoalex@xxxxxxxxx> wrote:
Hello Everyone,
Can someone help me to clarify this ?
I have a single-node 4.2.8 installation ( only two gluster storage domains - distributed  single drive volumes ). Now I just got two identintical servers and I would like to go for a 3 nodes bundle.
Is it possible ( after joining the new nodes to the cluster ) to expand the existing volumes across the new nodes and change them to replica 3 arbitrated ?
If so, could you share with me what would it be the procedure ?
Thank you very much !

Leo


--
Best regards, Leo David


--
Best regards, Leo David


--
Best regards, Leo David
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux