Re: Ceph cluster with 2 replicas

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



There are only two ways that size=2 can go:
A) You set min_size=1 and risk data loss
B) You set min_size=2 and your cluster stops every time you lose a
drive or reboot a machine

Neither of these are good options for most use cases; but there's
always an edge case. You should stay with size=3, min_size=2 unless
you have an unusual use case.

On Tue, Aug 17, 2021 at 10:33 AM Michel Niyoyita <micou12@xxxxxxxxx> wrote:
>
> Hi all ,
>
> Going to deploy a ceph cluster in production with  replicas size of 2 . Is
> there any inconvenience on the service side ?  I am going to change the
> default (3) to 2.
>
> Please advise.
>
> Regards.
>
> Michel
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux