Re: placing SSDs and SATAs pool in same hosts

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Jan 22, 2014 at 8:35 AM, zorg <zorg@xxxxxxxxxxxx> wrote:
> Hi,
> After reading the thread
> http://lists.ceph.com/pipermail/ceph-users-ceph.com/2013-June/002358.html
>
> We have done this crush map to make thing work.
> srv1 and srv1ssd are the same physical server (same srv2,3,4)
> we split it in the crush to make two parallel hierarchies.
> This example is working,
> I was just wondering if it's  the best way to achieve this

That looks good to me after a quick skim! You're splitting up your
SSDs into their own storage pool with their own rule separate from the
hard drives, and everything replicates across different hosts. I
assume that's what you're after.
-Greg
Software Engineer #42 @ http://inktank.com | http://ceph.com


>
> Thanks
>
>
> # begin crush map
>
> # devices
> device 0 osd.0
> device 1 osd.1
> device 2 osd.2
> device 3 osd.3
> device 4 osd.4
> device 5 osd.5
> device 6 osd.6
> device 7 osd.7
> device 8 osd.8
> device 9 osd.9
> device 10 osd.10
> device 11 osd.11
> device 12 osd.12
> device 13 osd.13
> device 14 osd.14
> device 15 osd.15
>
> # types
> type 0 osd
> type 1 host
> type 2 rack
> type 3 row
> type 4 room
> type 5 datacenter
> type 6 root
>
> # buckets
> host srv1 {
>     id -2        # do not change unnecessarily
>     # weight 5.690
>     alg straw
>     hash 0    # rjenkins1
>     item osd.1 weight 1.820
>     item osd.2 weight 1.820
>     item osd.3 weight 1.820
> }
> host srv2 {
>     id -3        # do not change unnecessarily
>     # weight 5.690
>     alg straw
>     hash 0    # rjenkins1
>     item osd.5 weight 1.820
>     item osd.14 weight 1.820
>     item osd.15 weight 1.820
> }
> host srv3 {
>     id -4        # do not change unnecessarily
>     # weight 5.690
>     alg straw
>     hash 0    # rjenkins1
>     item osd.7 weight 1.820
>     item osd.8 weight 1.820
>     item osd.9 weight 1.820
> }
> host srv4 {
>     id -5        # do not change unnecessarily
>     # weight 5.690
>     alg straw
>     hash 0    # rjenkins1
>     item osd.11 weight 1.820
>     item osd.12 weight 1.820
>     item osd.13 weight 1.820
> }
> host srv1ssd {
>     id -100
>     alg straw
>     hash 0
>     item osd.0 weight 0.230
> }
> host srv2ssd {
>     id -101
>     alg straw
>     hash 0
>     item osd.4 weight 0.230
> }
> host srv3ssd {
>     id -102
>     alg straw
>     hash 0
>     item osd.6 weight 0.230
> }
> host srv4ssd {
>     id -103
>     alg straw
>     hash 0
>     item osd.10 weight 0.230
> }
>
> root default {
>     id -1        # do not change unnecessarily
>     # weight 22.760
>     alg straw
>     hash 0    # rjenkins1
>     item srv1 weight 5.690
>     item srv2 weight 5.690
>     item srv3 weight 5.690
>     item srv4 weight 5.690
> }
>
> root ssd {
>     id -99
>     alg straw
>     hash 0
>     item srv1ssd weight 0.230
>     item srv2ssd weight 0.230
>     item srv3ssd weight 0.230
>     item srv4ssd weight 0.230
> }
>
>
> # rules
> rule data {
>     ruleset 0
>     type replicated
>     min_size 1
>     max_size 10
>     step take default
>     step chooseleaf firstn 0 type host
>     step emit
> }
> rule metadata {
>     ruleset 1
>     type replicated
>     min_size 1
>     max_size 10
>     step take default
>     step chooseleaf firstn 0 type host
>     step emit
> }
> rule sata {
>     ruleset 2
>     type replicated
>     min_size 1
>     max_size 10
>     step take default
>     step chooseleaf firstn 0 type host
>     step emit
> }
> rule ssd {
>     ruleset 3
>     type replicated
>     min_size 1
>     max_size 10
>     step take ssd
>     step chooseleaf firstn 0 type host
>     step emit
> }
>
> # end crush map
>
> cyril
>
> --
> probeSys - spécialiste GNU/Linux
> site web : http://www.probesys.com
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux