Hi,
this doesn't sound like a good idea: two hosts is usually a poor configuration for Ceph.Also, fewer disks on more servers is typically better than lots of disks in few servers.
But to answer your question: you could use a crush rule like this:
min_size 4
max_size 4
step take defaultstep choose firstn 2 type host
step choose firstn 2 type osd
step emit
2018-04-23 17:17 GMT+02:00 Christopher Meadors <christopher.meadors@xxxxxxxxxxxxxxxxxxxxx>:
I'm starting to get a small Ceph cluster running. I'm to the point where I've created a pool, and stored some test data in it, but I'm having trouble configuring the level of replication that I want.
The goal is to have two OSD host nodes, each with 20 OSDs. The target replication will be:
osd_pool_default_size = 4
osd_pool_default_min_size = 2
That is, I want two copies on each host, allowing for OSD failures or host failures without data loss.
How to best achieve this replication? Is this strictly a CRUSH map rule, or can it be done with the cluster conf? Pointers or examples would be greatly appreciated.
Thanks!
--
Chris
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
--
Paul Emmerich
Looking for help with your Ceph cluster? Contact us at https://croit.io
croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
Paul Emmerich
Looking for help with your Ceph cluster? Contact us at https://croit.io
croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com