I would suggest having some flash media for their own OSDs to put the cephfs metadata pool onto. That was a pretty significant boost for me when I moved the metadata pool onto flash media. My home setup is only 3 nodes and is running EC 2+1 on pure HDD OSDs with metadata on SSDs. It's been running stable and fine for a couple years now. I wouldn't suggest running EC 2+1 for any data you can't lose, but I can replace anything in there with some time.
On Wed, Aug 22, 2018 at 8:43 AM Paul Emmerich <paul.emmerich@xxxxxxxx> wrote:
Not 3+2, but we run 4+2, 6+2, 6+3, 5+3, and 8+3 with cephfs in
production. Most of them are HDDs without separate DB devices.
Paul
2018-08-22 14:27 GMT+02:00 Kevin Olbrich <ko@xxxxxxx>:
> Hi!
>
> I am in the progress of moving a local ("large", 24x1TB) ZFS RAIDZ2 to
> CephFS.
> This storage is used for backup images (large sequential reads and writes).
>
> To save space and have a RAIDZ2 (RAID6) like setup, I am planning the
> following profile:
>
> ceph osd erasure-code-profile set myprofile \
> k=3 \
> m=2 \
> ruleset-failure-domain=rack
>
> Performance is not the first priority, this is why I do not plan to
> outsource WAL/DB (broken NVMe = broken OSDs is more administrative overhead
> then single OSDs).
> Disks are attached by SAS multipath, throughput in general is no problem but
> I did not test with ceph yet.
>
> Is anyone using CephFS + bluestore + ec 3/2 + without WAL/DB-dev and is it
> working well?
>
> Thank you.
>
> Kevin
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
--
Paul Emmerich
Looking for help with your Ceph cluster? Contact us at https://croit.io
croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com