I guess it will be 2, since you did not hit a problem while taking out one rack. as soon as you take out an extra host after that, some PG's will have a size < 2, so the pool goes read-only and recovery should start. Once all pg's are back to size 2, the pool would be RW again.
To: "ceph-users" <ceph-users@xxxxxxxxxxxxxx>
Sent: Tuesday, October 23, 2018 6:44:52 PM
Subject: Crushmap and failure domains at rack level (ideally data-center level in the future)
Hello,
I want to create a crushmap rule where I can lose two racks of hosts and still be able to operate. I have tried the rule below, but it only allows me to operate (rados gateway) with one rack down and two racks up. If I lose any host in the two remaining racks my rados gateway stops responding. Here is my crushmap and rule. If anyone can point out what I am doing wrong it would be greatly appreciated. I’m very new to ceph so please forgive any incorrect terminology I have used.
# begin crush map
tunable choose_local_tries 0
tunable choose_local_fallback_tries 0
tunable choose_total_tries 50
tunable chooseleaf_descend_once 1
tunable chooseleaf_vary_r 1
tunable chooseleaf_stable 1
tunable straw_calc_version 1
tunable allowed_bucket_algs 54
# devices
device 0 osd.0 class hdd
device 1 osd.1 class hdd
device 2 osd.2 class hdd
device 3 osd.3 class hdd
device 4 osd.4 class hdd
device 5 osd.5 class hdd
device 6 osd.6 class hdd
device 7 osd.7 class hdd
device 8 osd.8 class hdd
device 9 osd.9 class hdd
device 10 osd.10 class hdd
device 11 osd.11 class hdd
device 12 osd.12 class hdd
device 13 osd.13 class hdd
device 14 osd.14 class hdd
device 15 osd.15 class hdd
device 16 osd.16 class hdd
device 17 osd.17 class hdd
device 18 osd.18 class hdd
device 19 osd.19 class hdd
device 20 osd.20 class hdd
device 21 osd.21 class hdd
device 22 osd.22 class hdd
device 23 osd.23 class hdd
device 24 osd.24 class hdd
device 25 osd.25 class hdd
device 26 osd.26 class hdd
device 27 osd.27 class hdd
device 28 osd.28 class hdd
device 29 osd.29 class hdd
device 30 osd.30 class hdd
device 31 osd.31 class hdd
device 32 osd.32 class hdd
device 33 osd.33 class hdd
device 34 osd.34 class hdd
device 35 osd.35 class hdd
device 36 osd.36 class hdd
device 37 osd.37 class hdd
device 38 osd.38 class hdd
device 39 osd.39 class hdd
device 40 osd.40 class hdd
device 41 osd.41 class hdd
device 42 osd.42 class hdd
device 43 osd.43 class hdd
device 44 osd.44 class hdd
device 45 osd.45 class hdd
device 46 osd.46 class hdd
device 47 osd.47 class hdd
device 48 osd.48 class hdd
device 49 osd.49 class hdd
device 50 osd.50 class hdd
device 51 osd.51 class hdd
device 52 osd.52 class hdd
device 53 osd.53 class hdd
device 54 osd.54 class hdd
device 55 osd.55 class hdd
device 56 osd.56 class hdd
device 57 osd.57 class hdd
device 58 osd.58 class hdd
device 59 osd.59 class hdd
device 60 osd.60 class hdd
device 61 osd.61 class hdd
device 62 osd.62 class hdd
device 63 osd.63 class hdd
device 64 osd.64 class hdd
device 65 osd.65 class hdd
device 66 osd.66 class hdd
device 67 osd.67 class hdd
device 68 osd.68 class hdd
device 69 osd.69 class hdd
device 70 osd.70 class hdd
device 71 osd.71 class hdd
device 72 osd.72 class hdd
device 73 osd.73 class hdd
device 74 osd.74 class hdd
device 75 osd.75 class hdd
device 76 osd.76 class hdd
device 77 osd.77 class hdd
device 78 osd.78 class hdd
device 79 osd.79 class hdd
device 80 osd.80 class hdd
device 81 osd.81 class hdd
device 82 osd.82 class hdd
device 83 osd.83 class hdd
device 84 osd.84 class hdd
device 85 osd.85 class hdd
device 86 osd.86 class hdd
device 87 osd.87 class hdd
device 88 osd.88 class hdd
device 89 osd.89 class hdd
device 90 osd.90 class hdd
device 91 osd.91 class hdd
device 92 osd.92 class hdd
device 93 osd.93 class hdd
device 94 osd.94 class hdd
device 95 osd.95 class hdd
device 96 osd.96 class hdd
device 97 osd.97 class hdd
device 98 osd.98 class hdd
device 99 osd.99 class hdd
device 100 osd.100 class hdd
device 101 osd.101 class hdd
device 102 osd.102 class hdd
device 103 osd.103 class hdd
device 104 osd.104 class hdd
device 105 osd.105 class hdd
device 106 osd.106 class hdd
device 107 osd.107 class hdd
device 108 osd.108 class hdd
device 109 osd.109 class hdd
device 110 osd.110 class hdd
device 111 osd.111 class hdd
device 112 osd.112 class hdd
device 113 osd.113 class hdd
device 114 osd.114 class hdd
device 115 osd.115 class hdd
device 116 osd.116 class hdd
device 117 osd.117 class hdd
device 118 osd.118 class hdd
device 119 osd.119 class hdd
device 120 osd.120 class hdd
device 121 osd.121 class hdd
device 122 osd.122 class hdd
device 123 osd.123 class hdd
device 124 osd.124 class hdd
device 125 osd.125 class hdd
device 126 osd.126 class hdd
device 127 osd.127 class hdd
device 128 osd.128 class hdd
device 129 osd.129 class hdd
device 130 osd.130 class hdd
device 131 osd.131 class hdd
device 132 osd.132 class hdd
device 133 osd.133 class hdd
device 134 osd.134 class hdd
device 135 osd.135 class hdd
device 136 osd.136 class hdd
device 137 osd.137 class hdd
device 138 osd.138 class hdd
device 139 osd.139 class hdd
device 140 osd.140 class hdd
device 141 osd.141 class hdd
device 142 osd.142 class hdd
device 143 osd.143 class hdd
# types
type 0 osd
type 1 host
type 2 chassis
type 3 rack
type 4 row
type 5 pdu
type 6 pod
type 7 room
type 8 datacenter
type 9 region
type 10 root
# buckets
host cephstorage02 {
id -3 # do not change unnecessarily
id -4 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.6 weight 7.277
item osd.10 weight 7.277
item osd.25 weight 7.277
item osd.34 weight 7.277
item osd.44 weight 7.277
item osd.55 weight 7.277
item osd.64 weight 7.277
item osd.75 weight 7.277
item osd.81 weight 7.277
item osd.95 weight 7.277
item osd.102 weight 7.277
item osd.108 weight 7.277
item osd.121 weight 7.277
item osd.126 weight 7.277
item osd.134 weight 7.277
item osd.138 weight 7.277
}
host cephstorage05 {
id -5 # do not change unnecessarily
id -6 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.3 weight 7.277
item osd.11 weight 7.277
item osd.23 weight 7.277
item osd.29 weight 7.277
item osd.40 weight 7.277
item osd.42 weight 7.277
item osd.51 weight 7.277
item osd.67 weight 7.277
item osd.79 weight 7.277
item osd.85 weight 7.277
item osd.90 weight 7.277
item osd.104 weight 7.277
item osd.113 weight 7.277
item osd.128 weight 7.277
item osd.140 weight 7.277
item osd.143 weight 7.277
}
host cephstorage06 {
id -7 # do not change unnecessarily
id -8 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.13 weight 7.277
item osd.18 weight 7.277
item osd.30 weight 7.277
item osd.38 weight 7.277
item osd.46 weight 7.277
item osd.48 weight 7.277
item osd.56 weight 7.277
item osd.59 weight 7.277
item osd.68 weight 7.277
item osd.80 weight 7.277
item osd.91 weight 7.277
item osd.101 weight 7.277
item osd.112 weight 7.277
item osd.120 weight 7.277
item osd.130 weight 7.277
item osd.141 weight 7.277
}
host cephstorage03 {
id -9 # do not change unnecessarily
id -10 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.2 weight 7.277
item osd.8 weight 7.277
item osd.21 weight 7.277
item osd.28 weight 7.277
item osd.36 weight 7.277
item osd.50 weight 7.277
item osd.62 weight 7.277
item osd.69 weight 7.277
item osd.72 weight 7.277
item osd.77 weight 7.277
item osd.83 weight 7.277
item osd.87 weight 7.277
item osd.103 weight 7.277
item osd.118 weight 7.277
item osd.129 weight 7.277
item osd.136 weight 7.277
}
host cephstorage04 {
id -11 # do not change unnecessarily
id -12 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.4 weight 7.277
item osd.14 weight 7.277
item osd.20 weight 7.277
item osd.27 weight 7.277
item osd.37 weight 7.277
item osd.45 weight 7.277
item osd.58 weight 7.277
item osd.65 weight 7.277
item osd.71 weight 7.277
item osd.82 weight 7.277
item osd.94 weight 7.277
item osd.98 weight 7.277
item osd.107 weight 7.277
item osd.119 weight 7.277
item osd.122 weight 7.277
item osd.131 weight 7.277
}
host cephstorage07 {
id -13 # do not change unnecessarily
id -14 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.0 weight 7.277
item osd.12 weight 7.277
item osd.22 weight 7.277
item osd.33 weight 7.277
item osd.43 weight 7.277
item osd.52 weight 7.277
item osd.57 weight 7.277
item osd.63 weight 7.277
item osd.70 weight 7.277
item osd.78 weight 7.277
item osd.88 weight 7.277
item osd.100 weight 7.277
item osd.110 weight 7.277
item osd.115 weight 7.277
item osd.123 weight 7.277
item osd.137 weight 7.277
}
host cephstorage01 {
id -15 # do not change unnecessarily
id -16 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.5 weight 7.277
item osd.15 weight 7.277
item osd.26 weight 7.277
item osd.35 weight 7.277
item osd.41 weight 7.277
item osd.54 weight 7.277
item osd.66 weight 7.277
item osd.74 weight 7.277
item osd.93 weight 7.277
item osd.99 weight 7.277
item osd.109 weight 7.277
item osd.114 weight 7.277
item osd.124 weight 7.277
item osd.132 weight 7.277
item osd.139 weight 7.277
item osd.142 weight 7.277
}
host cephstorage08 {
id -17 # do not change unnecessarily
id -18 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.1 weight 7.277
item osd.9 weight 7.277
item osd.16 weight 7.277
item osd.19 weight 7.277
item osd.32 weight 7.277
item osd.47 weight 7.277
item osd.49 weight 7.277
item osd.60 weight 7.277
item osd.73 weight 7.277
item osd.84 weight 7.277
item osd.89 weight 7.277
item osd.96 weight 7.277
item osd.105 weight 7.277
item osd.117 weight 7.277
item osd.127 weight 7.277
item osd.135 weight 7.277
}
host cephstorage09 {
id -19 # do not change unnecessarily
id -20 class hdd # do not change unnecessarily
# weight 116.437
alg straw2
hash 0 # rjenkins1
item osd.7 weight 7.277
item osd.17 weight 7.277
item osd.24 weight 7.277
item osd.31 weight 7.277
item osd.39 weight 7.277
item osd.53 weight 7.277
item osd.61 weight 7.277
item osd.76 weight 7.277
item osd.86 weight 7.277
item osd.92 weight 7.277
item osd.97 weight 7.277
item osd.106 weight 7.277
item osd.111 weight 7.277
item osd.116 weight 7.277
item osd.125 weight 7.277
item osd.133 weight 7.277
}
rack rack01 {
id -101
alg straw2
hash 0
item cephstorage01 weight 116.437
item cephstorage02 weight 116.437
item cephstorage03 weight 116.437
}
rack rack02 {
id -102
alg straw2
hash 0
item cephstorage04 weight 116.437
item cephstorage05 weight 116.437
item cephstorage06 weight 116.437
}
rack rack03 {
id -103
alg straw2
hash 0
item cephstorage07 weight 116.437
item cephstorage08 weight 116.437
item cephstorage09 weight 116.437
}
datacenter dc-1 {
id -201
alg straw2
item rack01 weight 349.311
item rack02 weight 349.311
item rack03 weight 349.311
}
region us-west {
id -301
alg straw2
item dc-1 weight 1047.933
}
root default {
id -1 # do not change unnecessarily
id -2 class hdd # do not change unnecessarily
# weight 1047.931
alg straw2
item wa-east weight 1047.933
}
# rules
rule replicated_rule {
id 0
type replicated
min_size 3
max_size 3
step take default
step choose firstn 0 type datacenter
step chooseleaf firstn 0 type rack
step emit
}
# end crush map
Dan Waterbly
| Senior Application Developer | 509.235.7500 x225 |
dan.waterbly@xxxxxxxxxx
WASHINGTON STATE ARCHIVES | DIGITAL ARCHIVES
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com