Re: Best osd scenario + ansible config?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

> So you need to think about failure domains.

Failure domains will be set to host.

> If you put all the DB's on one SSD and all the WAL's on another SSD then a failure of either of those SSD's will result in a failure of all the OSD's behind them. So in this case all 10 OSD's would have failed.
> 
> Splitting it to 5 OSD's you have RocksDb and WAL on each SSD this then results in a failure of an SSD only impacting 5 OSD's.
> 
> A failure of an SSD will take down all the OSD's that are behind that SSD.

That is what I wondered, thanks to confirm it.

> That's one of the reasons I would always say you need 1 nodes worth of spare capacity in the cluster to allow for automated re-builds to happen. 
> 
> As for your EC 7+5 I would have gone for some thing like 8+3 as then you have a spare node active in the cluster and can still provide full protection in the event of a failure of a node. 

Make sense! On another cluster, I have an EC 7+5 pool for cephfs but there are 4 servers per chassis. In case I lost one chassis, I still need
to access data. But for that cluster, you are right, 8+3 may be enough for redundancy.

> Think about software updates that require a reboot of a node. Any data written during that time will need recovering to bring it back to full protection where as if you have a spare node then that data could be written and not require a later recovery.

It is mostly a read-only cluster to distribute public datasets over S3 inside our network, it is fine for me if write operations are not fully
protected during a couple of days. All writes operations are managed by us to update datasets.

But as mentioned above, 8+3 may be a good compromise.

Best,

Yoann

> On 03/09/2019, 10:29, "Yoann Moulin" <yoann.moulin@xxxxxxx> wrote:
> 
>     Hello,
>     
>     I am deploying a new Nautilus cluster and I would like to know what would be the best OSD's scenario config in this case :
>     
>     10x 6TB Disk OSDs (data)
>      2x 480G SSD previously used for journal and can be used for WAL and/or DB
>     
>     Is it better to put all WAL on one SSD and all DBs on the other one? Or put WAL and DB of the first 5 OSDs on the first SSD and the 5 others on
>     the second one.
>     
>     A more general question, what is the impact on an OSD if we lose the WAL? The DB? Both?
>     
>     I plan to use EC 7+5 on 12 servers and I am OK if I lose one server temporarily. I have spare servers and I can easily add another one in this
>     cluster.
>     
>     To deploy this cluster, I use ceph-ansible (stable-4.0). I am not sure how to configure the playbook to use SSD and disks with LVM.
>     
>     https://github.com/ceph/ceph-ansible/blob/master/docs/source/osds/scenarios.rst
>     
>     Is this good?
>     
>     osd_objectstore: bluestore
>     lvm_volumes:
>       - data: data-lv1
>         data_vg: data-vg1
>         db: db-lv1
>         db_vg: db-vg1
>         wal: wal-lv1
>         wal_vg: wal-vg1
>       - data: data-lv2
>         data_vg: data-vg2
>         db: db-lv2
>         db_vg: db-vg2
>         wal: wal-lv2
>         wal_vg: wal-vg2
>     
>     
>     Is it possible to let the playbook configure LVM for each disk in a mixed case? It looks like I must configure LVM before running the playbook
>     but I am not sure if I missed something.
>     
>     Is wal_vg and db_vg can be identical (on VG per SSD shared with multiple OSDs)?
>     
>     Thanks for your help.
>     
>     Best regards,
>     
>     -- 
>     Yoann Moulin
>     EPFL IC-IT
>     _______________________________________________
>     ceph-users mailing list -- ceph-users@xxxxxxx
>     To unsubscribe send an email to ceph-users-leave@xxxxxxx
>     
>     
> 


-- 
Yoann Moulin
EPFL IC-IT
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux