Re: adding block.db to OSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Igore,
Am 27.04.20 um 15:03 schrieb Igor Fedotov:
> Just left a comment at https://tracker.ceph.com/issues/44509
> 
> Generally bdev-new-db performs no migration, RocksDB might eventually do
> that but no guarantee it moves everything.
> 
> One should use bluefs-bdev-migrate to do actual migration.
> 
> And I think that's the root cause for the above ticket.

perfect - this removed all spillover in seconds.

Greets,
Stefan


> Thanks,
> 
> Igor
> 
> On 4/24/2020 2:37 PM, Stefan Priebe - Profihost AG wrote:
>> No not a standalone Wal I wanted to ask whether bdev-new-db migrated
>> dB and Wal from hdd to ssd.
>>
>> Stefan
>>
>>> Am 24.04.2020 um 13:01 schrieb Igor Fedotov <ifedotov@xxxxxxx>:
>>>
>>> 
>>>
>>> Unless you have 3 different types of disks beyond OSD (e.g. HDD, SSD,
>>> NVMe) standalone WAL makes no sense.
>>>
>>>
>>> On 4/24/2020 1:58 PM, Stefan Priebe - Profihost AG wrote:
>>>> Is Wal device missing? Do I need to run *bluefs-bdev-new-db and Wal?*
>>>>
>>>> Greets,
>>>> Stefan
>>>>
>>>>> Am 24.04.2020 um 11:32 schrieb Stefan Priebe - Profihost AG
>>>>> <s.priebe@xxxxxxxxxxxx>:
>>>>>
>>>>> Hi Igor,
>>>>>
>>>>> there must be a difference. I purged osd.0 and recreated it.
>>>>>
>>>>> Now it gives:
>>>>> ceph tell osd.0 bench
>>>>> {
>>>>>    "bytes_written": 1073741824,
>>>>>    "blocksize": 4194304,
>>>>>    "elapsed_sec": 8.1554735639999993,
>>>>>    "bytes_per_sec": 131659040.46819863,
>>>>>    "iops": 31.389961354303033
>>>>> }
>>>>>
>>>>> What's wrong wiht adding a block.db device later?
>>>>>
>>>>> Stefan
>>>>>
>>>>> Am 23.04.20 um 20:34 schrieb Stefan Priebe - Profihost AG:
>>>>>> Hi,
>>>>>> if the OSDs are idle the difference is even more worse:
>>>>>> # ceph tell osd.0 bench
>>>>>> {
>>>>>>     "bytes_written": 1073741824,
>>>>>>     "blocksize": 4194304,
>>>>>>     "elapsed_sec": 15.396707875000001,
>>>>>>     "bytes_per_sec": 69738403.346825853,
>>>>>>     "iops": 16.626931034761871
>>>>>> }
>>>>>> # ceph tell osd.38 bench
>>>>>> {
>>>>>>     "bytes_written": 1073741824,
>>>>>>     "blocksize": 4194304,
>>>>>>     "elapsed_sec": 6.8903985170000004,
>>>>>>     "bytes_per_sec": 155831599.77624846,
>>>>>>     "iops": 37.153148597776521
>>>>>> }
>>>>>> Stefan
>>>>>> Am 23.04.20 um 14:39 schrieb Stefan Priebe - Profihost AG:
>>>>>>> Hi,
>>>>>>> Am 23.04.20 um 14:06 schrieb Igor Fedotov:
>>>>>>>> I don't recall any additional tuning to be applied to new DB
>>>>>>>> volume. And assume the hardware is pretty the same...
>>>>>>>>
>>>>>>>> Do you still have any significant amount of data spilled over
>>>>>>>> for these updated OSDs? If not I don't have any valid
>>>>>>>> explanation for the phenomena.
>>>>>>>
>>>>>>> just the 64k from here:
>>>>>>> https://tracker.ceph.com/issues/44509
>>>>>>>
>>>>>>>> You might want to try "ceph osd bench" to compare OSDs under
>>>>>>>> pretty the same load. Any difference observed
>>>>>>>
>>>>>>> Servers are the same HW. OSD Bench is:
>>>>>>> # ceph tell osd.0 bench
>>>>>>> {
>>>>>>>      "bytes_written": 1073741824,
>>>>>>>      "blocksize": 4194304,
>>>>>>>      "elapsed_sec": 16.091414781000001,
>>>>>>>      "bytes_per_sec": 66727620.822242722,
>>>>>>>      "iops": 15.909104543266945
>>>>>>> }
>>>>>>>
>>>>>>> # ceph tell osd.36 bench
>>>>>>> {
>>>>>>>      "bytes_written": 1073741824,
>>>>>>>      "blocksize": 4194304,
>>>>>>>      "elapsed_sec": 10.023828538,
>>>>>>>      "bytes_per_sec": 107118933.6419194,
>>>>>>>      "iops": 25.539143953780986
>>>>>>> }
>>>>>>>
>>>>>>>
>>>>>>> OSD 0 is a Toshiba MG07SCA12TA SAS 12G
>>>>>>> OSD 36 is a Seagate ST12000NM0008-2H SATA 6G
>>>>>>>
>>>>>>> SSDs are all the same like the rest of the HW. But both drives
>>>>>>> should give the same performance from their specs. The only other
>>>>>>> difference is that OSD 36 was directly created with the block.db
>>>>>>> device (Nautilus 14.2.7) and OSD 0 (14.2.8) does not.
>>>>>>>
>>>>>>> Stefan
>>>>>>>
>>>>>>>>
>>>>>>>> On 4/23/2020 8:35 AM, Stefan Priebe - Profihost AG wrote:
>>>>>>>>> Hello,
>>>>>>>>>
>>>>>>>>> is there anything else needed beside running:
>>>>>>>>> ceph-bluestore-tool --path /var/lib/ceph/osd/ceph-${OSD}
>>>>>>>>> bluefs-bdev-new-db --dev-target /dev/vgroup/lvdb-1
>>>>>>>>>
>>>>>>>>> I did so some weeks ago and currently i'm seeing that all osds
>>>>>>>>> originally deployed with --block-db show 10-20% I/O waits while
>>>>>>>>> all those got converted using ceph-bluestore-tool show 80-100%
>>>>>>>>> I/O waits.
>>>>>>>>>
>>>>>>>>> Also is there some tuning available to use more of the SSD? The
>>>>>>>>> SSD (block-db) is only saturated at 0-2%.
>>>>>>>>>
>>>>>>>>> Greets,
>>>>>>>>> Stefan
>>>>>>>>> _______________________________________________
>>>>>>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>>>>>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux