Am 11.01.2018 um 20:58 schrieb Sage Weil: > On Thu, 11 Jan 2018, Stefan Priebe - Profihost AG wrote: >> Hi Sage, >> >> this did not work like expected. I tested it in another smaller cluster >> and it resulted in about 6% misplaced objects. > > Can you narrow down at what stage the misplaced objects happened? ouch i saw this: # ceph balancer status { "active": true, "plans": [ "auto_2018-01-11_19:52:28" ], "mode": "crush-compat" } so might it be the balancer beeing executed while i was modifying the tree? Can i stop it and reexecute it manually? > > sage > >> >> Any ideas? >> >> Stefan >> Am 11.01.2018 um 08:09 schrieb Stefan Priebe - Profihost AG:HI >>> Thanks! Can this be done while still having jewel clients? >>> >>> Stefan >>> >>> Excuse my typo sent from my mobile phone. >>> >>> Am 10.01.2018 um 22:56 schrieb Sage Weil <sage@xxxxxxxxxxxx >>> <mailto:sage@xxxxxxxxxxxx>>: >>> >>>> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: >>>>> Am 10.01.2018 um 22:23 schrieb Sage Weil: >>>>>> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: >>>>>>> k, >>>>>>> >>>>>>> in the past we used the python crush optimize tool to reweight the osd >>>>>>> usage - it inserted a 2nd tree with $hostname-target-weight as >>>>>>> hostnames. >>>>>> >>>>>> Can you attach a 'ceph osd crush tree' (or partial output) so I can see >>>>>> what you mean? >>>>> >>>>> Sure - attached. >>>> >>>> Got it >>>> >>>>>>> Now the quesions are: >>>>>>> 1.) can we remove the tree? How? >>>>>>> 2.) Can we do this now or only after all clients are running Luminous? >>>>>>> 3.) is it enought to enable the mgr balancer module? >>>> >>>> First, >>>> >>>> ceph osd crush weight-set create-compat >>>> >>>> then for each osd, >>>> ceph osd crush weight-set reweight-compat <osd> <optimized-weight> >>>> ceph osd crush reweight <osd> <target-weight> >>>> >>>> That won't move any data but will keep your current optimized weights in >>>> the compat weight-set where they belong. >>>> >>>> Then you can remove the *-target-weight buckets. For each osd, >>>> >>>> ceph osd crush rm <osd> <ancestor>-target-weight >>>> >>>> and then for each remaining bucket >>>> >>>> ceph osd crush rm <foo>-target-weight >>>> >>>> Finally, turn on the balancer (or test it to see what it it wants to do >>>> with the optimize command.) >>>> >>>> HTH! >>>> sage >> -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html