Re: Question: CephFS + Bluestore

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I'm sorry I have mixed up some information. The actual ratio I have now is 0,0005% (100MB for 20TB data).


Regards,

Webert Lima
DevOps Engineer at MAV Tecnologia
Belo Horizonte - Brasil
IRC NICK - WebertRLZ

On Wed, May 9, 2018 at 11:32 AM, Webert de Souza Lima <webert.boss@xxxxxxxxx> wrote:
Hello,

Currently, I run Jewel + Filestore for cephfs, with SSD-only pools used for cephfs-metadata, and HDD-only pools for cephfs-data. The current metadata/data ratio is something like 0,25% (50GB metadata for 20TB data).

Regarding bluestore architecture, assuming I have: 

 - SSDs for WAL+DB
 - Spinning Disks for bluestore data.

would you recommend still store metadata in SSD-Only OSD nodes?
If not, is it recommended to dedicate some OSDs (Spindle+SSD for WAL/DB) for cephfs-metadata?

If I just have 2 pools (metadata and data) all sharing the same OSDs in the cluster, would it be enough for heavy-write cases?

Assuming min_size=2, size=3.

Thanks for your thoughts.

Regards,

Webert Lima
DevOps Engineer at MAV Tecnologia
Belo Horizonte - Brasil
IRC NICK - WebertRLZ

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux