Question: CephFS + Bluestore

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

Currently, I run Jewel + Filestore for cephfs, with SSD-only pools used for cephfs-metadata, and HDD-only pools for cephfs-data. The current metadata/data ratio is something like 0,25% (50GB metadata for 20TB data).

Regarding bluestore architecture, assuming I have: 

 - SSDs for WAL+DB
 - Spinning Disks for bluestore data.

would you recommend still store metadata in SSD-Only OSD nodes?
If not, is it recommended to dedicate some OSDs (Spindle+SSD for WAL/DB) for cephfs-metadata?

If I just have 2 pools (metadata and data) all sharing the same OSDs in the cluster, would it be enough for heavy-write cases?

Assuming min_size=2, size=3.

Thanks for your thoughts.

Regards,

Webert Lima
DevOps Engineer at MAV Tecnologia
Belo Horizonte - Brasil
IRC NICK - WebertRLZ
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux