Re: Customized Crush location hooks in Mimic

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I’m pretty sure the monitor command there won’t move intermediate buckets like the host. This is so if an osd has incomplete metadata it doesn’t inadvertently move 11 other OSDs into a different rack/row/whatever.

So in this case, it finds the host osd0001 and matches it, but since the crush map already knows about osd0001 it doesn’t pay any attention to the datacenter field.
Whereas if you tried setting it with mynewhost, the monitor wouldn’t know where that host exists and would look at the other fields to set it in the specified data center.
-Greg
On Fri, Nov 30, 2018 at 6:46 AM Oliver Freyermuth <freyermuth@xxxxxxxxxxxxxxxxxx> wrote:
Dear Cephalopodians,

sorry for the spam, but I found the following in mon logs just now and am finally out of ideas:
------------------------------------------------------------------------------------------
2018-11-30 15:43:05.207 7f9d64aac700  0 mon.mon001@0(leader) e3 handle_command mon_command({"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} v 0) v1
2018-11-30 15:43:05.207 7f9d64aac700  0 log_channel(audit) log [INF] : from='osd.1 10.160.12.101:6816/90528' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch
2018-11-30 15:43:05.208 7f9d64aac700  0 mon.mon001@0(leader) e3 handle_command mon_command({"prefix": "osd crush create-or-move", "id": 1, "weight":3.6824, "args": ["datacenter=FTD", "host=osd001", "root=default"]} v 0) v1
2018-11-30 15:43:05.208 7f9d64aac700  0 log_channel(audit) log [INF] : from='osd.1 10.160.12.101:6816/90528' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":3.6824, "args": ["datacenter=FTD", "host=osd001", "root=default"]}]: dispatch
2018-11-30 15:43:05.208 7f9d64aac700  0 mon.mon001@0(leader).osd e2464 create-or-move crush item name 'osd.1' initial_weight 3.6824 at location {datacenter=FTD,host=osd001,root=default}
------------------------------------------------------------------------------------------
So the request to move to datacenter=FTD arrives at the mon, but no action is taken, and the OSD is left in FTD_1.

Cheers,
        Oliver

Am 30.11.18 um 15:25 schrieb Oliver Freyermuth:
> Dear Cephalopodians,
>
> further experiments revealed that the crush-location-hook is indeed called!
> It's just my check (writing to a file in tmp from inside the hook) which somehow failed. Using "logger" works for debugging.
>
> So now, my hook outputs:
> host=osd001 datacenter=FTD root=default
> as explained before. I have also explicitly created the buckets beforehand in case that is needed.
>
> Tree looks like that:
> # ceph osd tree
> ID  CLASS WEIGHT   TYPE NAME            STATUS REWEIGHT PRI-AFF
>   -1       55.23582 root default
>   -9              0     datacenter FTD
> -12       18.41194     datacenter FTD_1
>   -3       18.41194         host osd001
>    0   hdd  3.68239             osd.0        up  1.00000 1.00000
>    1   hdd  3.68239             osd.1        up  1.00000 1.00000
>    2   hdd  3.68239             osd.2        up  1.00000 1.00000
>    3   hdd  3.68239             osd.3        up  1.00000 1.00000
>    4   hdd  3.68239             osd.4        up  1.00000 1.00000
> -11              0     datacenter FTD_2
>   -5       18.41194     host osd002
>    5   hdd  3.68239         osd.5            up  1.00000 1.00000
>    6   hdd  3.68239         osd.6            up  1.00000 1.00000
>    7   hdd  3.68239         osd.7            up  1.00000 1.00000
>    8   hdd  3.68239         osd.8            up  1.00000 1.00000
>    9   hdd  3.68239         osd.9            up  1.00000 1.00000
>   -7       18.41194     host osd003
>   10   hdd  3.68239         osd.10           up  1.00000 1.00000
>   11   hdd  3.68239         osd.11           up  1.00000 1.00000
>   12   hdd  3.68239         osd.12           up  1.00000 1.00000
>   13   hdd  3.68239         osd.13           up  1.00000 1.00000
>   14   hdd  3.68239         osd.14           up  1.00000 1.00000
>
> So naively, I would expect that when I restart osd.0, it should move itself into datacenter=FTD.
> But that does not happen...
>
> Any idea what I am missing?
>
> Cheers,
>      Oliver
>
>
>
> Am 30.11.18 um 11:44 schrieb Oliver Freyermuth:
>> Dear Cephalopodians,
>>
>> I'm probably missing something obvious, but I am at a loss here on how to actually make use of a customized crush location hook.
>>
>> I'm currently on "ceph version 13.2.1" on CentOS 7 (i.e. the last version before the upgrade-preventing bugs). Here's what I did:
>>
>> 1. Write a script /usr/local/bin/customized-ceph-crush-location. The script can be executed by user "ceph":
>>    # sudo -u ceph /usr/local/bin/customized-ceph-crush-location
>>    host=osd001 datacenter=FTD root=default
>>
>> 2. Add the following to ceph.conf:
>>   [osd]
>>   crush_location_hook = /usr/local/bin/customized-ceph-crush-location
>>
>> 3. Restart an OSD and confirm that is picked up:
>>   # systemctl restart ceph-osd@0
>>   # ceph config show-with-defaults osd.0
>>    ...
>>    crush_location_hook        /usr/local/bin/customized-ceph-crush-location  file
>>    ...
>>    osd_crush_update_on_start  true                                           default
>>    ...
>>
>> However, the script is not executed, and I can ensure that since the script should also write a log to /tmp, which is not created.
>> Also, the "datacenter" type does not show up in the crush tree.
>>
>> I have already disabled SELinux just to make sure.
>>
>> Any ideas what I am missing here?
>>
>> Cheers and thanks in advance,
>>      Oliver
>>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux