On Fri, Sep 06, 2013 at 08:21:07AM -0700, Sage Weil wrote: > On Fri, 6 Sep 2013, Chris Dunlop wrote: >> On Thu, Sep 05, 2013 at 07:55:52PM -0700, Sage Weil wrote: >>> Also, you should upgrade to dumpling. :) >> >> I've been considering it. It was initially a little scary with >> the various issues that were cropping up but that all seems to >> have quietened down. >> >> Of course I'd like my cluster to be clean before attempting an upgrade! > > Definitely. Let us know how it goes! :) Upgraded, directly from bobtail to dumpling. Well, that was a mite more traumatic than I expected. I had two issues, both my fault... Firstly, I didn't realise I should have restarted the osds one at a time rather than doing 'service ceph restart' on each host quickly in succession. Restarting them all at once meant everything was offline whilst "PGs are upgrading". Secondly, whilst I saw the 'osd crush update on start' issue in the release notes, and checked that my crush map hostnames match the actual hostnames, I have two separate pools (for fast SAS vs bulk SATA disks) and I stupidly only noticed the one which matched, but not the other which didn't match. So on restart all the osds moved into the one pool, and started rebalancing. The two issues at the same time produced quite the adrenaline rush! :-) My current crush configuration is below (host b2 is recently added and I haven't added it into the pools yet). Is there a better/recommended way of using the crush map to support separate pools to avoid setting 'osd crush update on start = false'? It doesn't seem that I can use the same 'host' names under the separate 'sas' and 'default' roots? Cheers, Chris ---------------------------------------------------------------------- # ceph osd tree # id weight type name up/down reweight -8 2 root sas -7 2 rack sas-rack-1 -5 1 host b4-sas 4 0.5 osd.4 up 1 5 0.5 osd.5 up 1 -6 1 host b5-sas 2 0.5 osd.2 up 1 3 0.5 osd.3 up 1 -1 12.66 root default -3 8 rack unknownrack -2 4 host b4 0 2 osd.0 up 1 7 2 osd.7 up 1 -4 4 host b5 1 2 osd.1 up 1 6 2 osd.6 up 1 -9 4.66 host b2 10 1.82 osd.10 up 1 11 1.82 osd.11 up 1 8 0.51 osd.8 up 1 9 0.51 osd.9 up 1 ---------------------------------------------------------------------- # begin crush map # devices device 0 osd.0 device 1 osd.1 device 2 osd.2 device 3 osd.3 device 4 osd.4 device 5 osd.5 device 6 osd.6 device 7 osd.7 device 8 osd.8 device 9 osd.9 device 10 osd.10 device 11 osd.11 # types type 0 osd type 1 host type 2 rack type 3 row type 4 room type 5 datacenter type 6 root # buckets host b4 { id -2 # do not change unnecessarily # weight 4.000 alg straw hash 0 # rjenkins1 item osd.0 weight 2.000 item osd.7 weight 2.000 } host b5 { id -4 # do not change unnecessarily # weight 4.000 alg straw hash 0 # rjenkins1 item osd.1 weight 2.000 item osd.6 weight 2.000 } rack unknownrack { id -3 # do not change unnecessarily # weight 8.000 alg straw hash 0 # rjenkins1 item b4 weight 4.000 item b5 weight 4.000 } host b2 { id -9 # do not change unnecessarily # weight 4.660 alg straw hash 0 # rjenkins1 item osd.10 weight 1.820 item osd.11 weight 1.820 item osd.8 weight 0.510 item osd.9 weight 0.510 } root default { id -1 # do not change unnecessarily # weight 12.660 alg straw hash 0 # rjenkins1 item unknownrack weight 8.000 item b2 weight 4.660 } host b4-sas { id -5 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.4 weight 0.500 item osd.5 weight 0.500 } host b5-sas { id -6 # do not change unnecessarily # weight 1.000 alg straw hash 0 # rjenkins1 item osd.2 weight 0.500 item osd.3 weight 0.500 } rack sas-rack-1 { id -7 # do not change unnecessarily # weight 2.000 alg straw hash 0 # rjenkins1 item b4-sas weight 1.000 item b5-sas weight 1.000 } root sas { id -8 # do not change unnecessarily # weight 2.000 alg straw hash 0 # rjenkins1 item sas-rack-1 weight 2.000 } # rules rule data { ruleset 0 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } rule metadata { ruleset 1 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } rule rbd { ruleset 2 type replicated min_size 1 max_size 10 step take default step chooseleaf firstn 0 type host step emit } rule rbd-sas { ruleset 3 type replicated min_size 1 max_size 10 step take sas step chooseleaf firstn 0 type host step emit } # end crush map ---------------------------------------------------------------------- -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html