Thanks! Can this be done while still having jewel clients? Stefan Excuse my typo sent from my mobile phone. Am 10.01.2018 um 22:56 schrieb Sage Weil <sage@xxxxxxxxxxxx <mailto:sage@xxxxxxxxxxxx>>: > On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: >> Am 10.01.2018 um 22:23 schrieb Sage Weil: >>> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: >>>> k, >>>> >>>> in the past we used the python crush optimize tool to reweight the osd >>>> usage - it inserted a 2nd tree with $hostname-target-weight as >>>> hostnames. >>> >>> Can you attach a 'ceph osd crush tree' (or partial output) so I can see >>> what you mean? >> >> Sure - attached. > > Got it > >>>> Now the quesions are: >>>> 1.) can we remove the tree? How? >>>> 2.) Can we do this now or only after all clients are running Luminous? >>>> 3.) is it enought to enable the mgr balancer module? > > First, > > ceph osd crush weight-set create-compat > > then for each osd, > ceph osd crush weight-set reweight-compat <osd> <optimized-weight> > ceph osd crush reweight <osd> <target-weight> > > That won't move any data but will keep your current optimized weights in > the compat weight-set where they belong. > > Then you can remove the *-target-weight buckets. For each osd, > > ceph osd crush rm <osd> <ancestor>-target-weight > > and then for each remaining bucket > > ceph osd crush rm <foo>-target-weight > > Finally, turn on the balancer (or test it to see what it it wants to do > with the optimize command.) > > HTH! > sage -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html