Re: Best osd scenario + ansible config?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

> Just a note:
> With 7+5 you will need 13 host for access your data in case one goes down.

As far as I know, EC 7+5 imply ‘erasure size 12 min_size 8’. So I need at least 8 servers access to my data,

k=7, m=5,
size = k+m = 12
min_size = k+1 = 8

Am I wrong?

> Expected in the nexts version allow access data with the EC numbers.

I think it is still possible to set min_size = k in Nautilus but it is not recommended.

Best,

Yoann

> -----Mensaje original-----
> De: Yoann Moulin <yoann.moulin@xxxxxxx> 
> Enviado el: martes, 3 de septiembre de 2019 11:28
> Para: ceph-users@xxxxxxx
> Asunto:  Best osd scenario + ansible config?
> 
> Hello,
> 
> I am deploying a new Nautilus cluster and I would like to know what would be
> the best OSD's scenario config in this case :
> 
> 10x 6TB Disk OSDs (data)
>  2x 480G SSD previously used for journal and can be used for WAL and/or DB
> 
> Is it better to put all WAL on one SSD and all DBs on the other one? Or put
> WAL and DB of the first 5 OSDs on the first SSD and the 5 others on the
> second one.
> 
> A more general question, what is the impact on an OSD if we lose the WAL?
> The DB? Both?
> 
> I plan to use EC 7+5 on 12 servers and I am OK if I lose one server
> temporarily. I have spare servers and I can easily add another one in this
> cluster.
> 
> To deploy this cluster, I use ceph-ansible (stable-4.0). I am not sure how
> to configure the playbook to use SSD and disks with LVM.
> 
> https://github.com/ceph/ceph-ansible/blob/master/docs/source/osds/scenarios.
> rst
> 
> Is this good?
> 
> osd_objectstore: bluestore
> lvm_volumes:
>   - data: data-lv1
>     data_vg: data-vg1
>     db: db-lv1
>     db_vg: db-vg1
>     wal: wal-lv1
>     wal_vg: wal-vg1
>   - data: data-lv2
>     data_vg: data-vg2
>     db: db-lv2
>     db_vg: db-vg2
>     wal: wal-lv2
>     wal_vg: wal-vg2
> 
> 
> Is it possible to let the playbook configure LVM for each disk in a mixed
> case? It looks like I must configure LVM before running the playbook but I
> am not sure if I missed something.
> 
> Is wal_vg and db_vg can be identical (on VG per SSD shared with multiple
> OSDs)?
> 
> Thanks for your help.
> 
> Best regards,
> 
> --
> Yoann Moulin
> EPFL IC-IT
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email
> to ceph-users-leave@xxxxxxx
> 


-- 
Yoann Moulin
EPFL IC-IT
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux