Re: Stretch cluster questions

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, May 4, 2022 at 1:25 AM Eneko Lacunza <elacunza@xxxxxxxxx> wrote:

> Hi Gregory,
>
> El 3/5/22 a las 22:30, Gregory Farnum escribió:
>
> On Mon, Apr 25, 2022 at 12:57 AM Eneko Lacunza <elacunza@xxxxxxxxx> <elacunza@xxxxxxxxx> wrote:
>
> We're looking to deploy a stretch cluster for a 2-CPD deployment. I have
> read the following docs:https://docs.ceph.com/en/latest/rados/operations/stretch-mode/#stretch-clusters
>
> I have some questions:
>
> - Can we have multiple pools in a stretch cluster?
>
> Yes.
>
>
> - Can we have multiple different crush rules in a stretch cluster? I'm
> asking this because the command for stretch mode activation asks for a
> rule...
>
> Right, so what happens there is any pool with a default rule gets
> switched to the specified CRUSH rule. That doesn't stop you from
> changing the rule after engaging stretch mode, or giving it a
> non-default rule ahead of time. You just have to be careful to make
> sure it satisfies the stretch mode rules about placing across data
> centers.
>
> So the only purpose of the "stretch_rule" param to "enable_stretch_mode"
> is replacing default replicated rule?
>
> If so, when stretch mode is activated, the described special behaviour
> applies to all crush rules/pools:
>
> - the OSDs will only take PGs active when they peer across data centers
> (or whatever other CRUSH bucket type you specified), assuming both are alive
>
> - Pools will increase in size from the default 3 to 4, expecting 2 copies
> in each site (for size=2 pools will it increase to 4 too?)
>
> Is this accurate?
>

Yep! AFAIK most clusters just use the default pool for replication, barring
mixed media types. *shrug*


>
> We want to have different purpose pools on this Ceph cluster:
>
> - Important VM disks, with 2 copies in each DC (SSD class)
> - Ephemeral VM disks, with just 2 copies overall (SSD class)
> - Backup data in just one DC (HDD class).
>
> Objective of the 2-DC deployment is disaster recovery, HA isn't
> required, but I'll take it if deployment is reasonable :-) .
>
> I'm leery of this for the reasons described in the docs — if you don't
> have 2 replicas per site, you lose data availability every time an OSD
> goes down for any reason (or else you have a window while recovery
> happens where the data is not physically available in both sites,
> which rather negates the purpose).
>
>
> Is this because the following: "the OSDs will only take PGs active when
> they peer across data centers (or whatever other CRUSH bucket type you
> specified), assuming both are alive"?
>

>
> This conclusion wasn't obvious to me, but after your reply and a third
> read, now seems expected :-)
>

Right!


>
> Thanks for your comments. I'll be doing some tests, will write back if I
> find something unexpected (to me?) :-)
>
> Cheers
>
> Eneko Lacunza
> Zuzendari teknikoa | Director técnico
> Binovo IT Human Project
>
> Tel. +34 943 569 206 | https://www.binovo.esAstigarragako Bidea, 2 - 2º <https://www.google.com/maps/search/Astigarragako+Bidea,+2+-+2%C2%BA?entry=gmail&source=g> izda. Oficina 10-11, 20180 Oiartzun
> https://www.youtube.com/user/CANALBINOVOhttps://www.linkedin.com/company/37269706/
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux