Thanks a lot guys.
Best,
German
2016-03-24 15:55 GMT-03:00 Sean Redmond <sean.redmond1@xxxxxxxxx>:
Hi German,For Data to be split over the racks you should set the crush rule set to 'step chooseleaf firstn 0 type rack' instead of 'step chooseleaf firstn 0 type host'ThanksOn Wed, Mar 23, 2016 at 3:50 PM, German Anders <ganders@xxxxxxxxxxxx> wrote:_______________________________________________Hi all,I had a question, I'm in the middle of a new ceph deploy cluster and I've 6 OSD servers between two racks, so rack1 would have osdserver1,3 and 5, and rack2 osdserver2,4 and 6. I've edited the following crush map and I want to know if it's ok and also if the objects would be stored one on each rack-host. So, if I lost one rack, I had one copy on the other rack/server:
http://pastebin.com/raw/QJf1VeeJAlso If I need to run any command in order to 'apply' the new crush map to the existing pools (actually only two):- 0 rbd (pg_num: 4096 | pgp_num: 4096 | size: 2 | min_size: 1)- 1 cinder-volumes (pg_num: 4096 | pgp_num: 4096 | size: 2 | min_size: 1)
# ceph --cluster cephIB osd tree
ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY
-1 51.29668 root default
-8 26.00958 rack cage5-rack1
-2 8.66986 host cibn01
0 0.72249 osd.0 up 1.00000 1.00000
1 0.72249 osd.1 up 1.00000 1.00000
2 0.72249 osd.2 up 1.00000 1.00000
3 0.72249 osd.3 up 1.00000 1.00000
4 0.72249 osd.4 up 1.00000 1.00000
5 0.72249 osd.5 up 1.00000 1.00000
6 0.72249 osd.6 up 1.00000 1.00000
7 0.72249 osd.7 up 1.00000 1.00000
8 0.72249 osd.8 up 1.00000 1.00000
9 0.72249 osd.9 up 1.00000 1.00000
10 0.72249 osd.10 up 1.00000 1.00000
11 0.72249 osd.11 up 1.00000 1.00000
-4 8.66986 host cibn03
24 0.72249 osd.24 up 1.00000 1.00000
25 0.72249 osd.25 up 1.00000 1.00000
26 0.72249 osd.26 up 1.00000 1.00000
27 0.72249 osd.27 up 1.00000 1.00000
28 0.72249 osd.28 up 1.00000 1.00000
29 0.72249 osd.29 up 1.00000 1.00000
30 0.72249 osd.30 up 1.00000 1.00000
31 0.72249 osd.31 up 1.00000 1.00000
32 0.72249 osd.32 up 1.00000 1.00000
33 0.72249 osd.33 up 1.00000 1.00000
34 0.72249 osd.34 up 1.00000 1.00000
35 0.72249 osd.35 up 1.00000 1.00000
-6 8.66986 host cibn05
48 0.72249 osd.48 up 1.00000 1.00000
49 0.72249 osd.49 up 1.00000 1.00000
50 0.72249 osd.50 up 1.00000 1.00000
51 0.72249 osd.51 up 1.00000 1.00000
52 0.72249 osd.52 up 1.00000 1.00000
53 0.72249 osd.53 up 1.00000 1.00000
54 0.72249 osd.54 up 1.00000 1.00000
55 0.72249 osd.55 up 1.00000 1.00000
56 0.72249 osd.56 up 1.00000 1.00000
57 0.72249 osd.57 up 1.00000 1.00000
58 0.72249 osd.58 up 1.00000 1.00000
59 0.72249 osd.59 up 1.00000 1.00000
-9 25.28709 rack cage5-rack2
-3 8.66986 host cibn02
12 0.72249 osd.12 up 1.00000 1.00000
13 0.72249 osd.13 up 1.00000 1.00000
14 0.72249 osd.14 up 1.00000 1.00000
15 0.72249 osd.15 up 1.00000 1.00000
16 0.72249 osd.16 up 1.00000 1.00000
17 0.72249 osd.17 up 1.00000 1.00000
18 0.72249 osd.18 up 1.00000 1.00000
19 0.72249 osd.19 up 1.00000 1.00000
20 0.72249 osd.20 up 1.00000 1.00000
21 0.72249 osd.21 up 1.00000 1.00000
22 0.72249 osd.22 up 1.00000 1.00000
23 0.72249 osd.23 up 1.00000 1.00000
-5 8.66986 host cibn04
36 0.72249 osd.36 up 1.00000 1.00000
37 0.72249 osd.37 up 1.00000 1.00000
38 0.72249 osd.38 up 1.00000 1.00000
39 0.72249 osd.39 up 1.00000 1.00000
40 0.72249 osd.40 up 1.00000 1.00000
41 0.72249 osd.41 up 1.00000 1.00000
42 0.72249 osd.42 up 1.00000 1.00000
43 0.72249 osd.43 up 1.00000 1.00000
44 0.72249 osd.44 up 1.00000 1.00000
45 0.72249 osd.45 up 1.00000 1.00000
46 0.72249 osd.46 up 1.00000 1.00000
47 0.72249 osd.47 up 1.00000 1.00000
-7 7.94737 host cibn06
60 0.72249 osd.60 up 1.00000 1.00000
61 0.72249 osd.61 up 1.00000 1.00000
62 0.72249 osd.62 up 1.00000 1.00000
63 0.72249 osd.63 up 1.00000 1.00000
64 0.72249 osd.64 up 1.00000 1.00000
65 0.72249 osd.65 up 1.00000 1.00000
66 0.72249 osd.66 up 1.00000 1.00000
67 0.72249 osd.67 up 1.00000 1.00000
68 0.72249 osd.68 up 1.00000 1.00000
69 0.72249 osd.69 up 1.00000 1.00000
70 0.72249 osd.70 up 1.00000 1.00000Ceph version 9.2.1 (752b6a3020c3de74e07d2a8b4c5e48dab5a6b6fd)Thanks in advance,Best,German
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com