Hi,
I'm running a 4 node test 'cluster' (VMs on my workstation) that I've
upgraded to Luminous RC. Specifically I wanted to test having each node
with 1 spinning device and one solid state state so I could try out
device classes to create fast and slow(er) pools.
I started with 4 filestore osds (comimg from the Jewel pre-upgrade), and
added 4 more, all of which were Bluestore on the ssds.
I used crushtool to set the device classes (see crush test diff below).
That all went very smoothly, with only a couple of things that seemed
weird. Firstly the crush/osd tree output is a bit strange (but I could
get to the point where it make sense):
$ sudo ceph osd tree
ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY
-15 0.23196 root default~ssd
-11 0.05699 host ceph1~ssd
4 0.05699 osd.4 up 1.00000 1.00000
-12 0.05899 host ceph2~ssd
5 0.05899 osd.5 up 1.00000 1.00000
-13 0.05699 host ceph3~ssd
6 0.05699 osd.6 up 1.00000 1.00000
-14 0.05899 host ceph4~ssd
7 0.05899 osd.7 up 1.00000 1.00000
-10 0.07996 root default~hdd
-6 0.01999 host ceph1~hdd
0 0.01999 osd.0 up 1.00000 1.00000
-7 0.01999 host ceph2~hdd
1 0.01999 osd.1 up 1.00000 1.00000
-8 0.01999 host ceph3~hdd
2 0.01999 osd.2 up 1.00000 1.00000
-9 0.01999 host ceph4~hdd
3 0.01999 osd.3 up 1.00000 1.00000
-1 0.31198 root default
-2 0.07700 host ceph1
0 0.01999 osd.0 up 1.00000 1.00000
4 0.05699 osd.4 up 1.00000 1.00000
-3 0.07899 host ceph2
1 0.01999 osd.1 up 1.00000 1.00000
5 0.05899 osd.5 up 1.00000 1.00000
-4 0.07700 host ceph3
2 0.01999 osd.2 up 1.00000 1.00000
6 0.05699 osd.6 up 1.00000 1.00000
-5 0.07899 host ceph4
3 0.01999 osd.3 up 1.00000 1.00000
7 0.05899 osd.7 up 1.00000 1.00000
But the osd df output is baffling, I've got two identical lines for each
osd (hard to see immediately - sorting by osd id would make it easier).
This is not ideal, particularly as for the bluestore guys there is no
other way to work out utilization. Any ideas - have I done something
obviously wrong here that is triggering the 2 lines?
$ sudo ceph osd df
ID WEIGHT REWEIGHT SIZE USE AVAIL %USE VAR PGS
4 0.05699 1.00000 60314M 1093M 59221M 1.81 1.27 0
5 0.05899 1.00000 61586M 1234M 60351M 2.00 1.40 0
6 0.05699 1.00000 60314M 1248M 59066M 2.07 1.45 0
7 0.05899 1.00000 61586M 1209M 60376M 1.96 1.37 0
0 0.01999 1.00000 25586M 43812k 25543M 0.17 0.12 45
1 0.01999 1.00000 25586M 42636k 25544M 0.16 0.11 37
2 0.01999 1.00000 25586M 44336k 25543M 0.17 0.12 53
3 0.01999 1.00000 25586M 42716k 25544M 0.16 0.11 57
0 0.01999 1.00000 25586M 43812k 25543M 0.17 0.12 45
4 0.05699 1.00000 60314M 1093M 59221M 1.81 1.27 0
1 0.01999 1.00000 25586M 42636k 25544M 0.16 0.11 37
5 0.05899 1.00000 61586M 1234M 60351M 2.00 1.40 0
2 0.01999 1.00000 25586M 44336k 25543M 0.17 0.12 53
6 0.05699 1.00000 60314M 1248M 59066M 2.07 1.45 0
3 0.01999 1.00000 25586M 42716k 25544M 0.16 0.11 57
7 0.05899 1.00000 61586M 1209M 60376M 1.96 1.37 0
TOTAL 338G 4955M 333G 1.43
MIN/MAX VAR: 0.11/1.45 STDDEV: 0.97
The modifications to crush map
--- crush.txt.orig 2017-06-28 14:38:38.067669000 +1200
+++ crush.txt 2017-06-28 14:41:22.071669000 +1200
@@ -8,14 +8,14 @@
tunable allowed_bucket_algs 54
# devices
-device 0 osd.0
-device 1 osd.1
-device 2 osd.2
-device 3 osd.3
-device 4 osd.4
-device 5 osd.5
-device 6 osd.6
-device 7 osd.7
+device 0 osd.0 class hdd
+device 1 osd.1 class hdd
+device 2 osd.2 class hdd
+device 3 osd.3 class hdd
+device 4 osd.4 class ssd
+device 5 osd.5 class ssd
+device 6 osd.6 class ssd
+device 7 osd.7 class ssd
# types
type 0 osd
@@ -80,7 +80,7 @@
type replicated
min_size 1
max_size 10
- step take default
+ step take default class hdd
step chooseleaf firstn 0 type host
step emit
}
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html