Re: Luminous - replace old target-weight tree from osdmap with mgr balancer

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am 11.01.2018 um 20:58 schrieb Sage Weil:
> On Thu, 11 Jan 2018, Stefan Priebe - Profihost AG wrote:
>> Hi Sage,
>>
>> this did not work like expected. I tested it in another smaller cluster
>> and it resulted in about 6% misplaced objects.
> 
> Can you narrow down at what stage the misplaced objects happened?

ouch i saw this:
# ceph balancer status
{
    "active": true,
    "plans": [
        "auto_2018-01-11_19:52:28"
    ],
    "mode": "crush-compat"
}

so might it be the balancer beeing executed while i was modifying the tree?
Can i stop it and reexecute it manually?

> 
> sage
> 
>>
>> Any ideas?
>>
>> Stefan
>> Am 11.01.2018 um 08:09 schrieb Stefan Priebe - Profihost AG:HI
>>> Thanks! Can this be done while still having jewel clients?
>>>
>>> Stefan
>>>
>>> Excuse my typo sent from my mobile phone.
>>>
>>> Am 10.01.2018 um 22:56 schrieb Sage Weil <sage@xxxxxxxxxxxx
>>> <mailto:sage@xxxxxxxxxxxx>>:
>>>
>>>> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote:
>>>>> Am 10.01.2018 um 22:23 schrieb Sage Weil:
>>>>>> On Wed, 10 Jan 2018, Stefan Priebe - Profihost AG wrote:
>>>>>>> k,
>>>>>>>
>>>>>>> in the past we used the python crush optimize tool to reweight the osd
>>>>>>> usage - it inserted a 2nd tree with $hostname-target-weight as
>>>>>>> hostnames.
>>>>>>
>>>>>> Can you attach a 'ceph osd crush tree' (or partial output) so I can see
>>>>>> what you mean?
>>>>>
>>>>> Sure - attached.
>>>>
>>>> Got it
>>>>
>>>>>>> Now the quesions are:
>>>>>>> 1.) can we remove the tree? How?
>>>>>>> 2.) Can we do this now or only after all clients are running Luminous?
>>>>>>> 3.) is it enought to enable the mgr balancer module?
>>>>
>>>> First,
>>>>
>>>> ceph osd crush weight-set create-compat
>>>>
>>>> then for each osd,
>>>> ceph osd crush weight-set reweight-compat <osd> <optimized-weight>
>>>> ceph osd crush reweight <osd> <target-weight>
>>>>
>>>> That won't move any data but will keep your current optimized weights in
>>>> the compat weight-set where they belong.
>>>>
>>>> Then you can remove the *-target-weight buckets.  For each osd,
>>>>
>>>> ceph osd crush rm <osd> <ancestor>-target-weight
>>>>
>>>> and then for each remaining bucket
>>>>
>>>> ceph osd crush rm <foo>-target-weight
>>>>
>>>> Finally, turn on the balancer (or test it to see what it it wants to do
>>>> with the optimize command.)
>>>>
>>>> HTH!
>>>> sage
>>
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux