Hi Sage, this did not work like expected. I tested it in another smaller cluster and it resulted in about 6% misplaced objects. Any ideas? Stefan Am 11.01.2018 um 08:09 schrieb Stefan Priebe - Profihost AG:HI > Thanks! Can this be done while still having jewel clients? > > Stefan > > Excuse my typo sent from my mobile phone. > > Am 10.01.2018 um 22:56 schrieb Sage Weil <sage@xxxxxxxxxxxx > <mailto:sage@xxxxxxxxxxxx>>: > >> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: >>> Am 10.01.2018 um 22:23 schrieb Sage Weil: >>>> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: >>>>> k, >>>>> >>>>> in the past we used the python crush optimize tool to reweight the osd >>>>> usage - it inserted a 2nd tree with $hostname-target-weight as >>>>> hostnames. >>>> >>>> Can you attach a 'ceph osd crush tree' (or partial output) so I can see >>>> what you mean? >>> >>> Sure - attached. >> >> Got it >> >>>>> Now the quesions are: >>>>> 1.) can we remove the tree? How? >>>>> 2.) Can we do this now or only after all clients are running Luminous? >>>>> 3.) is it enought to enable the mgr balancer module? >> >> First, >> >> ceph osd crush weight-set create-compat >> >> then for each osd, >> ceph osd crush weight-set reweight-compat <osd> <optimized-weight> >> ceph osd crush reweight <osd> <target-weight> >> >> That won't move any data but will keep your current optimized weights in >> the compat weight-set where they belong. >> >> Then you can remove the *-target-weight buckets. For each osd, >> >> ceph osd crush rm <osd> <ancestor>-target-weight >> >> and then for each remaining bucket >> >> ceph osd crush rm <foo>-target-weight >> >> Finally, turn on the balancer (or test it to see what it it wants to do >> with the optimize command.) >> >> HTH! >> sage -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html