3) Forgive the style, it'll be going into the docs shortly :) It's possible to have multiple independent crush heirarchies within the same crush map. Suppose you want to have pools default to osds backed by large spinning disks but have some pools mapped to osds backed by fast ssds: device 0 osd.0 device 1 osd.1 device 2 osd.2 device 3 osd.3 device 4 osd.4 device 5 osd.5 device 6 osd.6 device 7 osd.7 host ceph-osd-ssd-server-1 { id -1 alg straw hash 0 item osd.0 weight 1.00 item osd.1 weight 1.00 } host ceph-osd-ssd-server-2 { id -2 alg straw hash 0 item osd.2 weight 1.00 item osd.3 weight 1.00 } host ceph-osd-platter-server-1 { id -3 alg straw hash 0 item osd.4 weight 1.00 item osd.5 weight 1.00 } host ceph-osd-platter-server-2 { id -4 alg straw hash 0 item osd.6 weight 1.00 item osd.7 weight 1.00 } root platter { id -5 alg straw hash 0 item ceph-osd-platter-server-1 weight 2.00 item ceph-osd-platter-server-2 weight 2.00 } root ssd { id -6 alg straw hash 0 item ceph-osd-ssd-server-1 weight 2.00 item ceph-osd-ssd-server-2 weight 2.00 } rule data { ruleset 0 type replicated min_size 2 max_size 2 step take platter step chooseleaf 0 type host step emit } rule metadata { ruleset 1 type replicated min_size 0 max_size 10 step take platter step chooseleaf 0 type host step emit } rule rbd { ruleset 2 type replicated min_size 0 max_size 10 step take platter step chooseleaf 0 type host step emit } rule platter { ruleset 3 type replicated min_size 0 max_size 10 step take platter step chooseleaf 0 type host step emit } rule ssd { ruleset 4 type replicated min_size 0 max_size 10 step take ssd step chooseleaf 0 type host step emit } rule ssd-primary { ruleset 4 type replicated min_size 0 max_size 10 step take ssd step chooseleaf 1 type host step emit step take platter step chooseleaf -1 type host step emit } You can then set a pool to use the ssd rule by: ceph osd pool set <poolname> crush_ruleset 4 Similarly, using the ssd-primary rule will cause each pg in the pool to be placed with an ssd as the primary and platters as the replicas. -Sam On Mon, Dec 10, 2012 at 11:17 AM, Alexandre Maumené <alexandre@xxxxxxxxxxx> wrote: > Hello all, > > I have a few questions about Ceph: > > 1) Is it possible to run a cluster with "some" lantecy between monitor > nodes? Latency will be 30ms at worst. > > 2) When using RBD what are the best practices for a direct mount using > XFS filesystem? And for a qemu/kvm devices? I'm thinking about > writeback, rbd_cache, ... > > 3) About the CRUSH map, how can I separate 2 pools on different OSD? > I'd like to setup a cluster with different disks (like SATA/SAS) and I > want to be able to specify on which disks (or OSD) my data are going > to be write. > > Thanks in advance for any answer. > > Regards, > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html