Re: device class : nvme

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The nvme devices show up as ssd, so I have to manually reclassify them on my cluster. 

Sent from my iPhone. Typos are Apple's fault. 

> On Nov 21, 2019, at 4:10 PM, Kyle Bader <kyle.bader@xxxxxxxxx> wrote:
> 
> We ssd device class on rook-ceph built clusters on m5 instances
> (devices appear as nvme)
> 
>> On Thu, Nov 21, 2019 at 2:48 PM Mark Nelson <mnelson@xxxxxxxxxx> wrote:
>> 
>> 
>>> On 11/21/19 4:46 PM, Mark Nelson wrote:
>>> On 11/21/19 4:25 PM, Sage Weil wrote:
>>>> Adding dev@xxxxxxx
>>>> 
>>>> On Thu, 21 Nov 2019, Muhammad Ahmad wrote:
>>>>> While trying to research how crush maps are used/modified I stumbled
>>>>> upon these device classes.
>>>>> https://ceph.io/community/new-luminous-crush-device-classes/
>>>>> 
>>>>> I wanted to highlight that having nvme as a separate class will
>>>>> eventually break and should be removed.
>>>>> 
>>>>> There is already a push within the industry to consolidate future
>>>>> command sets and NVMe will likely be it. In other words, NVMe HDDs are
>>>>> not too far off. In fact, the recent October OCP F2F discussed this
>>>>> topic in detail.
>>>>> 
>>>>> If the classification is based on performance then command set
>>>>> (SATA/SAS/NVMe) is probably not the right classification.
>>>> I opened a PR that does this:
>>>> 
>>>>    https://github.com/ceph/ceph/pull/31796
>>>> 
>>>> I can't remember seeing 'nvme' as a device class on any real cluster;
>>>> the
>>>> exceptoin is my basement one, and I think the only reason it ended up
>>>> that
>>>> way was because I deployed bluestore *very* early on (with ceph-disk)
>>>> and
>>>> the is_nvme() detection helper doesn't work with LVM.  That's my
>>>> theory at
>>>> least.. can anybody with bluestore on NVMe devices confirm? Does anybody
>>>> see class 'nvme' devices in their cluster?
>>>> 
>>>> Thanks!
>>>> sage
>>>> 
>>> 
>>> Here's what we've got on the new performance nodes with Intel NVMe
>>> drives:
>>> 
>>> 
>>> ID  CLASS WEIGHT   TYPE NAME
>>> -1       64.00000 root default
>>> -3       64.00000     rack localrack
>>> -2        8.00000         host o03
>>>  0   ssd  1.00000             osd.0
>>>  1   ssd  1.00000             osd.1
>>>  2   ssd  1.00000             osd.2
>>>  3   ssd  1.00000             osd.3
>>>  4   ssd  1.00000             osd.4
>>>  5   ssd  1.00000             osd.5
>>>  6   ssd  1.00000             osd.6
>>>  7   ssd  1.00000             osd.7
>>> 
>>> 
>>> Mark
>>> 
>> 
>> I should probably clarify that this cluster was built with cbt though!
>> 
>> 
>> Mark
>> _______________________________________________
>> Dev mailing list -- dev@xxxxxxx
>> To unsubscribe send an email to dev-leave@xxxxxxx
_______________________________________________
Dev mailing list -- dev@xxxxxxx
To unsubscribe send an email to dev-leave@xxxxxxx




[Index of Archives]     [CEPH Users]     [Ceph Devel]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux