On Thu, 11 Jan 2018, Stefan Priebe - Profihost AG wrote: > Hi Sage, > > this did not work like expected. I tested it in another smaller cluster > and it resulted in about 6% misplaced objects. Can you narrow down at what stage the misplaced objects happened? sage > > Any ideas? > > Stefan > Am 11.01.2018 um 08:09 schrieb Stefan Priebe - Profihost AG:HI > > Thanks! Can this be done while still having jewel clients? > > > > Stefan > > > > Excuse my typo sent from my mobile phone. > > > > Am 10.01.2018 um 22:56 schrieb Sage Weil <sage@xxxxxxxxxxxx > > <mailto:sage@xxxxxxxxxxxx>>: > > > >> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: > >>> Am 10.01.2018 um 22:23 schrieb Sage Weil: > >>>> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote: > >>>>> k, > >>>>> > >>>>> in the past we used the python crush optimize tool to reweight the osd > >>>>> usage - it inserted a 2nd tree with $hostname-target-weight as > >>>>> hostnames. > >>>> > >>>> Can you attach a 'ceph osd crush tree' (or partial output) so I can see > >>>> what you mean? > >>> > >>> Sure - attached. > >> > >> Got it > >> > >>>>> Now the quesions are: > >>>>> 1.) can we remove the tree? How? > >>>>> 2.) Can we do this now or only after all clients are running Luminous? > >>>>> 3.) is it enought to enable the mgr balancer module? > >> > >> First, > >> > >> ceph osd crush weight-set create-compat > >> > >> then for each osd, > >> ceph osd crush weight-set reweight-compat <osd> <optimized-weight> > >> ceph osd crush reweight <osd> <target-weight> > >> > >> That won't move any data but will keep your current optimized weights in > >> the compat weight-set where they belong. > >> > >> Then you can remove the *-target-weight buckets. For each osd, > >> > >> ceph osd crush rm <osd> <ancestor>-target-weight > >> > >> and then for each remaining bucket > >> > >> ceph osd crush rm <foo>-target-weight > >> > >> Finally, turn on the balancer (or test it to see what it it wants to do > >> with the optimize command.) > >> > >> HTH! > >> sage > >