On Wed, Jan 22, 2014 at 8:35 AM, zorg <zorg@xxxxxxxxxxxx> wrote: > Hi, > After reading the thread > http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-June/002358.html > > We have done this crush map to make thing work. > srv1 and srv1ssd are the same physical server (same srv2,3,4) > we split it in the crush to make two parallel hierarchies. > This example is working, > I was just wondering if it's the best way to achieve this That looks good to me after a quick skim! You're splitting up your SSDs into their own storage pool with their own rule separate from the hard drives, and everything replicates across different hosts. I assume that's what you're after. -Greg Software Engineer #42 @ http://inktank.com | http://ceph.com > > Thanks > > > # begin crush map > > # devices > device 0 osd.0 > device 1 osd.1 > device 2 osd.2 > device 3 osd.3 > device 4 osd.4 > device 5 osd.5 > device 6 osd.6 > device 7 osd.7 > device 8 osd.8 > device 9 osd.9 > device 10 osd.10 > device 11 osd.11 > device 12 osd.12 > device 13 osd.13 > device 14 osd.14 > device 15 osd.15 > > # types > type 0 osd > type 1 host > type 2 rack > type 3 row > type 4 room > type 5 datacenter > type 6 root > > # buckets > host srv1 { > id -2 # do not change unnecessarily > # weight 5.690 > alg straw > hash 0 # rjenkins1 > item osd.1 weight 1.820 > item osd.2 weight 1.820 > item osd.3 weight 1.820 > } > host srv2 { > id -3 # do not change unnecessarily > # weight 5.690 > alg straw > hash 0 # rjenkins1 > item osd.5 weight 1.820 > item osd.14 weight 1.820 > item osd.15 weight 1.820 > } > host srv3 { > id -4 # do not change unnecessarily > # weight 5.690 > alg straw > hash 0 # rjenkins1 > item osd.7 weight 1.820 > item osd.8 weight 1.820 > item osd.9 weight 1.820 > } > host srv4 { > id -5 # do not change unnecessarily > # weight 5.690 > alg straw > hash 0 # rjenkins1 > item osd.11 weight 1.820 > item osd.12 weight 1.820 > item osd.13 weight 1.820 > } > host srv1ssd { > id -100 > alg straw > hash 0 > item osd.0 weight 0.230 > } > host srv2ssd { > id -101 > alg straw > hash 0 > item osd.4 weight 0.230 > } > host srv3ssd { > id -102 > alg straw > hash 0 > item osd.6 weight 0.230 > } > host srv4ssd { > id -103 > alg straw > hash 0 > item osd.10 weight 0.230 > } > > root default { > id -1 # do not change unnecessarily > # weight 22.760 > alg straw > hash 0 # rjenkins1 > item srv1 weight 5.690 > item srv2 weight 5.690 > item srv3 weight 5.690 > item srv4 weight 5.690 > } > > root ssd { > id -99 > alg straw > hash 0 > item srv1ssd weight 0.230 > item srv2ssd weight 0.230 > item srv3ssd weight 0.230 > item srv4ssd weight 0.230 > } > > > # rules > rule data { > ruleset 0 > type replicated > min_size 1 > max_size 10 > step take default > step chooseleaf firstn 0 type host > step emit > } > rule metadata { > ruleset 1 > type replicated > min_size 1 > max_size 10 > step take default > step chooseleaf firstn 0 type host > step emit > } > rule sata { > ruleset 2 > type replicated > min_size 1 > max_size 10 > step take default > step chooseleaf firstn 0 type host > step emit > } > rule ssd { > ruleset 3 > type replicated > min_size 1 > max_size 10 > step take ssd > step chooseleaf firstn 0 type host > step emit > } > > # end crush map > > cyril > > -- > probeSys - spécialiste GNU/Linux > site web : http://www.probesys.com > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com