Re: Crushmap and failure domains at rack level (ideally data-center level in the future)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I think I was able to figure out what happened. Some of my rgw pools had a min_size of 1 and some had a min_size of 2.

I am not certain how this happened but my theory is I must have had at least one of the pools created before I modified the crushmap rules. After the rule was modified new pools must have been created. Would this cause what I witnessed?

At any rate, I changed all pools that were using this specific crushrule’s min_size to 1. Killed two racks of nodes and the rados gateway kept chugging along.

Thank you for your pointer leading me to the solution.

-Dan




On Thu, Oct 25, 2018 at 9:12 PM -0700, "Waterbly, Dan" <dan.waterbly@xxxxxxxxxx> wrote:

Sorry for the delay but ceph osd pool get <poolname> min_size returns 1. Where does this value come from?




On Tue, Oct 23, 2018 at 5:47 PM -0700, "Bastiaan Visser" <bvisser@xxxxxxxxxx> wrote:

Something must be wrong, since you have min_size 3 the pool should go read only once you take out the first rack. Probably even when you take out the first host.

What is the outputput of ceph osd pool get <poolname> min_size ?

I guess it will be 2, since you did not hit a problem while taking out one rack. as soon as you take out an extra host after that, some PG's will have a size < 2, so the pool goes read-only and recovery should start. Once all pg's are back to size 2, the pool would be RW again.


From: "Waterbly, Dan" <dan.waterbly@xxxxxxxxxx>
To: "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
Sent: Tuesday, October 23, 2018 6:44:52 PM
Subject: [ceph-users] Crushmap and failure domains at rack level (ideally data-center level in the future)

Hello,

 

I want to create a crushmap rule where I can lose two racks of hosts and still be able to operate. I have tried the rule below, but it only allows me to operate (rados gateway) with one rack down and two racks up. If I lose any host in the two remaining racks my rados gateway stops responding. Here is my crushmap and rule. If anyone can point out what I am doing wrong it would be greatly appreciated. I’m very new to ceph so please forgive any incorrect terminology I have used.

 

# begin crush map

tunable choose_local_tries 0

tunable choose_local_fallback_tries 0

tunable choose_total_tries 50

tunable chooseleaf_descend_once 1

tunable chooseleaf_vary_r 1

tunable chooseleaf_stable 1

tunable straw_calc_version 1

tunable allowed_bucket_algs 54

 

# devices

device 0 osd.0 class hdd

device 1 osd.1 class hdd

device 2 osd.2 class hdd

device 3 osd.3 class hdd

device 4 osd.4 class hdd

device 5 osd.5 class hdd

device 6 osd.6 class hdd

device 7 osd.7 class hdd

device 8 osd.8 class hdd

device 9 osd.9 class hdd

device 10 osd.10 class hdd

device 11 osd.11 class hdd

device 12 osd.12 class hdd

device 13 osd.13 class hdd

device 14 osd.14 class hdd

device 15 osd.15 class hdd

device 16 osd.16 class hdd

device 17 osd.17 class hdd

device 18 osd.18 class hdd

device 19 osd.19 class hdd

device 20 osd.20 class hdd

device 21 osd.21 class hdd

device 22 osd.22 class hdd

device 23 osd.23 class hdd

device 24 osd.24 class hdd

device 25 osd.25 class hdd

device 26 osd.26 class hdd

device 27 osd.27 class hdd

device 28 osd.28 class hdd

device 29 osd.29 class hdd

device 30 osd.30 class hdd

device 31 osd.31 class hdd

device 32 osd.32 class hdd

device 33 osd.33 class hdd

device 34 osd.34 class hdd

device 35 osd.35 class hdd

device 36 osd.36 class hdd

device 37 osd.37 class hdd

device 38 osd.38 class hdd

device 39 osd.39 class hdd

device 40 osd.40 class hdd

device 41 osd.41 class hdd

device 42 osd.42 class hdd

device 43 osd.43 class hdd

device 44 osd.44 class hdd

device 45 osd.45 class hdd

device 46 osd.46 class hdd

device 47 osd.47 class hdd

device 48 osd.48 class hdd

device 49 osd.49 class hdd

device 50 osd.50 class hdd

device 51 osd.51 class hdd

device 52 osd.52 class hdd

device 53 osd.53 class hdd

device 54 osd.54 class hdd

device 55 osd.55 class hdd

device 56 osd.56 class hdd

device 57 osd.57 class hdd

device 58 osd.58 class hdd

device 59 osd.59 class hdd

device 60 osd.60 class hdd

device 61 osd.61 class hdd

device 62 osd.62 class hdd

device 63 osd.63 class hdd

device 64 osd.64 class hdd

device 65 osd.65 class hdd

device 66 osd.66 class hdd

device 67 osd.67 class hdd

device 68 osd.68 class hdd

device 69 osd.69 class hdd

device 70 osd.70 class hdd

device 71 osd.71 class hdd

device 72 osd.72 class hdd

device 73 osd.73 class hdd

device 74 osd.74 class hdd

device 75 osd.75 class hdd

device 76 osd.76 class hdd

device 77 osd.77 class hdd

device 78 osd.78 class hdd

device 79 osd.79 class hdd

device 80 osd.80 class hdd

device 81 osd.81 class hdd

device 82 osd.82 class hdd

device 83 osd.83 class hdd

device 84 osd.84 class hdd

device 85 osd.85 class hdd

device 86 osd.86 class hdd

device 87 osd.87 class hdd

device 88 osd.88 class hdd

device 89 osd.89 class hdd

device 90 osd.90 class hdd

device 91 osd.91 class hdd

device 92 osd.92 class hdd

device 93 osd.93 class hdd

device 94 osd.94 class hdd

device 95 osd.95 class hdd

device 96 osd.96 class hdd

device 97 osd.97 class hdd

device 98 osd.98 class hdd

device 99 osd.99 class hdd

device 100 osd.100 class hdd

device 101 osd.101 class hdd

device 102 osd.102 class hdd

device 103 osd.103 class hdd

device 104 osd.104 class hdd

device 105 osd.105 class hdd

device 106 osd.106 class hdd

device 107 osd.107 class hdd

device 108 osd.108 class hdd

device 109 osd.109 class hdd

device 110 osd.110 class hdd

device 111 osd.111 class hdd

device 112 osd.112 class hdd

device 113 osd.113 class hdd

device 114 osd.114 class hdd

device 115 osd.115 class hdd

device 116 osd.116 class hdd

device 117 osd.117 class hdd

device 118 osd.118 class hdd

device 119 osd.119 class hdd

device 120 osd.120 class hdd

device 121 osd.121 class hdd

device 122 osd.122 class hdd

device 123 osd.123 class hdd

device 124 osd.124 class hdd

device 125 osd.125 class hdd

device 126 osd.126 class hdd

device 127 osd.127 class hdd

device 128 osd.128 class hdd

device 129 osd.129 class hdd

device 130 osd.130 class hdd

device 131 osd.131 class hdd

device 132 osd.132 class hdd

device 133 osd.133 class hdd

device 134 osd.134 class hdd

device 135 osd.135 class hdd

device 136 osd.136 class hdd

device 137 osd.137 class hdd

device 138 osd.138 class hdd

device 139 osd.139 class hdd

device 140 osd.140 class hdd

device 141 osd.141 class hdd

device 142 osd.142 class hdd

device 143 osd.143 class hdd

 

# types

type 0 osd

type 1 host

type 2 chassis

type 3 rack

type 4 row

type 5 pdu

type 6 pod

type 7 room

type 8 datacenter

type 9 region

type 10 root

 

# buckets

host cephstorage02 {

        id -3           # do not change unnecessarily

        id -4 class hdd         # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.6 weight 7.277

        item osd.10 weight 7.277

        item osd.25 weight 7.277

        item osd.34 weight 7.277

        item osd.44 weight 7.277

        item osd.55 weight 7.277

        item osd.64 weight 7.277

        item osd.75 weight 7.277

        item osd.81 weight 7.277

        item osd.95 weight 7.277

        item osd.102 weight 7.277

        item osd.108 weight 7.277

        item osd.121 weight 7.277

        item osd.126 weight 7.277

        item osd.134 weight 7.277

        item osd.138 weight 7.277

}

host cephstorage05 {

        id -5           # do not change unnecessarily

        id -6 class hdd         # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.3 weight 7.277

        item osd.11 weight 7.277

        item osd.23 weight 7.277

        item osd.29 weight 7.277

        item osd.40 weight 7.277

        item osd.42 weight 7.277

        item osd.51 weight 7.277

        item osd.67 weight 7.277

        item osd.79 weight 7.277

        item osd.85 weight 7.277

        item osd.90 weight 7.277

        item osd.104 weight 7.277

        item osd.113 weight 7.277

        item osd.128 weight 7.277

        item osd.140 weight 7.277

        item osd.143 weight 7.277

}

host cephstorage06 {

        id -7           # do not change unnecessarily

        id -8 class hdd         # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.13 weight 7.277

        item osd.18 weight 7.277

        item osd.30 weight 7.277

        item osd.38 weight 7.277

        item osd.46 weight 7.277

        item osd.48 weight 7.277

        item osd.56 weight 7.277

        item osd.59 weight 7.277

        item osd.68 weight 7.277

        item osd.80 weight 7.277

        item osd.91 weight 7.277

        item osd.101 weight 7.277

        item osd.112 weight 7.277

        item osd.120 weight 7.277

        item osd.130 weight 7.277

        item osd.141 weight 7.277

}

host cephstorage03 {

        id -9           # do not change unnecessarily

        id -10 class hdd                # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.2 weight 7.277

        item osd.8 weight 7.277

        item osd.21 weight 7.277

        item osd.28 weight 7.277

        item osd.36 weight 7.277

        item osd.50 weight 7.277

        item osd.62 weight 7.277

        item osd.69 weight 7.277

        item osd.72 weight 7.277

        item osd.77 weight 7.277

        item osd.83 weight 7.277

        item osd.87 weight 7.277

        item osd.103 weight 7.277

        item osd.118 weight 7.277

        item osd.129 weight 7.277

        item osd.136 weight 7.277

}

host cephstorage04 {

        id -11          # do not change unnecessarily

        id -12 class hdd                # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.4 weight 7.277

        item osd.14 weight 7.277

        item osd.20 weight 7.277

        item osd.27 weight 7.277

        item osd.37 weight 7.277

        item osd.45 weight 7.277

        item osd.58 weight 7.277

        item osd.65 weight 7.277

        item osd.71 weight 7.277

        item osd.82 weight 7.277

        item osd.94 weight 7.277

        item osd.98 weight 7.277

        item osd.107 weight 7.277

        item osd.119 weight 7.277

        item osd.122 weight 7.277

        item osd.131 weight 7.277

}

host cephstorage07 {

        id -13          # do not change unnecessarily

        id -14 class hdd                # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.0 weight 7.277

        item osd.12 weight 7.277

        item osd.22 weight 7.277

        item osd.33 weight 7.277

        item osd.43 weight 7.277

        item osd.52 weight 7.277

        item osd.57 weight 7.277

        item osd.63 weight 7.277

        item osd.70 weight 7.277

        item osd.78 weight 7.277

        item osd.88 weight 7.277

        item osd.100 weight 7.277

        item osd.110 weight 7.277

        item osd.115 weight 7.277

        item osd.123 weight 7.277

        item osd.137 weight 7.277

}

host cephstorage01 {

        id -15          # do not change unnecessarily

        id -16 class hdd                # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.5 weight 7.277

        item osd.15 weight 7.277

        item osd.26 weight 7.277

        item osd.35 weight 7.277

        item osd.41 weight 7.277

        item osd.54 weight 7.277

        item osd.66 weight 7.277

        item osd.74 weight 7.277

        item osd.93 weight 7.277

        item osd.99 weight 7.277

        item osd.109 weight 7.277

        item osd.114 weight 7.277

        item osd.124 weight 7.277

        item osd.132 weight 7.277

        item osd.139 weight 7.277

        item osd.142 weight 7.277

}

host cephstorage08 {

        id -17          # do not change unnecessarily

        id -18 class hdd                # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.1 weight 7.277

        item osd.9 weight 7.277

        item osd.16 weight 7.277

        item osd.19 weight 7.277

        item osd.32 weight 7.277

        item osd.47 weight 7.277

        item osd.49 weight 7.277

        item osd.60 weight 7.277

        item osd.73 weight 7.277

        item osd.84 weight 7.277

        item osd.89 weight 7.277

        item osd.96 weight 7.277

        item osd.105 weight 7.277

        item osd.117 weight 7.277

        item osd.127 weight 7.277

        item osd.135 weight 7.277

}

host cephstorage09 {

        id -19          # do not change unnecessarily

        id -20 class hdd                # do not change unnecessarily

        # weight 116.437

        alg straw2

        hash 0  # rjenkins1

        item osd.7 weight 7.277

        item osd.17 weight 7.277

        item osd.24 weight 7.277

        item osd.31 weight 7.277

        item osd.39 weight 7.277

        item osd.53 weight 7.277

        item osd.61 weight 7.277

        item osd.76 weight 7.277

        item osd.86 weight 7.277

        item osd.92 weight 7.277

        item osd.97 weight 7.277

        item osd.106 weight 7.277

        item osd.111 weight 7.277

        item osd.116 weight 7.277

        item osd.125 weight 7.277

        item osd.133 weight 7.277

}

 

rack rack01 {

        id -101

        alg straw2

        hash 0

        item cephstorage01 weight 116.437

        item cephstorage02 weight 116.437

        item cephstorage03 weight 116.437

}

 

rack rack02 {

        id -102

        alg straw2

        hash 0

        item cephstorage04 weight 116.437

        item cephstorage05 weight 116.437

        item cephstorage06 weight 116.437

}

 

rack rack03 {

        id -103

        alg straw2

        hash 0

        item cephstorage07 weight 116.437

        item cephstorage08 weight 116.437

        item cephstorage09 weight 116.437

}

 

datacenter dc-1 {

        id -201

        alg straw2

        item rack01 weight 349.311

        item rack02 weight 349.311

        item rack03 weight 349.311

}

 

region us-west {

        id -301

        alg straw2

        item dc-1 weight 1047.933

}

 

root default {

        id -1           # do not change unnecessarily

        id -2 class hdd         # do not change unnecessarily

        # weight 1047.931

        alg straw2

        item wa-east weight 1047.933

}

 

# rules

rule replicated_rule {

        id 0

        type replicated

        min_size 3

        max_size 3

        step take default

        step choose firstn 0 type datacenter

        step chooseleaf firstn 0 type rack

        step emit

}

 

# end crush map

 

Dan Waterbly | Senior Application Developer | 509.235.7500 x225 | dan.waterbly@xxxxxxxxxx

WASHINGTON STATE ARCHIVES | DIGITAL ARCHIVES

 


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux