Re: Introduce flash OSD's to Nautilus installation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



... unless you use the reclassify tooling:

https://docs.ceph.com/en/latest/rados/operations/crush-map-edits/#migrating-from-a-legacy-ssd-rule-to-device-classes


On Thu, 17 Sep 2020, 19:30 Dan van der Ster, <dan@xxxxxxxxxxxxxx> wrote:

> Hi,
>
> AFAIR the device types adds a bunch of shadow devices e.g. osd.1~hdd or
> something like that... And those shadow devs have a different crush id than
> the original, untyped device.
>
> So, alas, I don't think your test is complete, and yes I expect that your
> data would move if you change the rule properly.
>
> Cheers, Dan
>
>
>
> On Thu, 17 Sep 2020, 18:36 Mathias Lindberg, <mathlin@xxxxxxxxxxx> wrote:
>
>> Hi,
>>
>> We have a 1.2PB Nautilus installation primarily using CephFS for our
>> HPC-resources.
>> Our OSD’s have spinning disks and NvME devices for WAL and DB in an
>> LVM-setup.
>>
>> The CephFS metadata pool resides on spinning disks, and I wonder if there
>> is any point from a performance perspective to put that on flash?
>> Trying to google that does not provide for a single straight answer, some
>> say its is heavily cached on the MDS and does not benefit that much from
>> flash while others argue it is significantly faster putting it on flash.
>>
>> We would regardless like to introduce flash OSD’s.
>> In order to do that and not have data moving on to the ssd’s OSD’s we add
>> we need add crush map rules that place data on ssd and hdd exclusively .
>>
>> But from reading previous posts to the list adding rules like that would
>> trigger PG’s to start migrating.
>>
>> But when I (reluctantly) manually edit the crush map, adding a new class
>> ssd (using a previously unused id) and adding the new rules I need
>> crushtool seems to think no data vill shuffle and that maps are equivalent.
>> Is it that simple, or am i doing it wrong?
>> I would have preferred to have used some sort of tool like the crushtool
>> reclassify (if applicable) to make these changes, but I can’t get the hang
>> of that at all.
>> The plan is to then change the crush rule for the exiting pools to the
>> new ones with device classes.
>>
>> Thank you for any pointers or advice.
>>
>> [root@cephyr-mon1 crushtest]# diff -u crush_comp crush_comp_corr
>> --- crush_comp 2020-09-17 17:11:37.125310334 +0200
>> +++ crush_comp_corr 2020-09-17 17:14:28.022704876 +0200
>> @@ -574,6 +574,7 @@
>>  root default {
>>   id -1 # do not change unnecessarily
>>   id -2 class hdd # do not change unnecessarily
>> +        id -29 class ssd
>>   # weight 2087.180
>>   alg straw2
>>   hash 0 # rjenkins1
>> @@ -710,5 +711,25 @@
>>   step chooseleaf firstn 0 type host
>>   step emit
>>  }
>> +rule replicated_ssd {
>> +        id 15
>> +        type replicated
>> +        min_size 1
>> +        max_size 10
>> +        step take default class ssd
>> +        step chooseleaf firstn 0 type host
>> +        step emit
>> +}
>> +rule cephyrfs_data_hdd {
>> +        id 16
>> +        type erasure
>> +        min_size 3
>> +        max_size 12
>> +        step set_chooseleaf_tries 5
>> +        step set_choose_tries 100
>> +        step take default class hdd
>> +        step chooseleaf indep 0 type host
>> +        step emit
>> +}
>>
>>  # end crush map
>>
>> [root@cephyr-mon1 crushtest]# crushtool -i crush_comp.c --compare
>> crush_comp_corr.c
>> rule 0 had 0/10240 mismatched mappings (0)
>> rule 1 had 0/6144 mismatched mappings (0)
>> rule 6 had 0/10240 mismatched mappings (0)
>> rule 7 had 0/4096 mismatched mappings (0)
>> rule 8 had 0/4096 mismatched mappings (0)
>> rule 9 had 0/4096 mismatched mappings (0)
>> rule 10 had 0/4096 mismatched mappings (0)
>> rule 11 had 0/4096 mismatched mappings (0)
>> rule 12 had 0/4096 mismatched mappings (0)
>> rule 13 had 0/4096 mismatched mappings (0)
>> rule 14 had 0/10240 mismatched mappings (0)
>> maps appear equivalent
>>
>> Regards,
>> Mathias Lindberg
>>
>> Tel: +46 (0)31 7723059
>> Mob: +46 (0)723 526107
>> Mathias Lindberg
>> mathlin@xxxxxxxxxxx
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux