Re: cephadm to setup wal/db on nvme

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




> Thank you for reply,
> 
> I have created two class SSD and NvME and assigned them to crush maps.

You don't have enough drives to keep them separate.  Set the NVMe drives back to "ssd" and just make one pool.

> 
> $ ceph osd crush rule ls
> replicated_rule
> ssd_pool
> nvme_pool
> 
> 
> Running benchmarks on nvme is the worst performing. SSD showing much better
> results compared to NvME.

You have more SATA SSDs and thus more OSDs, than NVMe SSDs.


> NvME model is Samsung_SSD_980_PRO_1TB

Client-grade, don't expect much from it.


> 
> #### NvME pool benchmark with 3x replication
> 
> # rados -p test-nvme -t 64 -b 4096 bench 10 write
> hints = 1
> Maintaining 64 concurrent writes of 4096 bytes to objects of size 4096 for
> up to 10 seconds or 0 objects
> Object prefix: benchmark_data_os-ctrl1_1931595
>  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg
> lat(s)
>    0       0         0         0         0         0           -
> 0
>    1      64      5541      5477   21.3917   21.3945   0.0134898
> 0.0116529
>    2      64     11209     11145   21.7641   22.1406  0.00939951
> 0.0114506
>    3      64     17036     16972   22.0956   22.7617  0.00938263
> 0.0112938
>    4      64     23187     23123   22.5776   24.0273  0.00863939
> 0.0110473
>    5      64     29753     29689   23.1911   25.6484  0.00925603
> 0.0107662
>    6      64     36222     36158   23.5369   25.2695   0.0100759
> 0.010606
>    7      63     42997     42934   23.9551   26.4688  0.00902186
> 0.0104246
>    8      64     49859     49795   24.3102   26.8008  0.00884379
> 0.0102765
>    9      64     56429     56365   24.4601   25.6641  0.00989885
> 0.0102124
>   10      31     62727     62696   24.4869   24.7305   0.0115833
> 0.0102027
> Total time run:         10.0064
> Total writes made:      62727
> Write size:             4096
> Object size:            4096
> Bandwidth (MB/sec):     24.4871
> Stddev Bandwidth:       1.85423
> Max bandwidth (MB/sec): 26.8008       <------------   Only 26MB/s for nvme
> disk
> Min bandwidth (MB/sec): 21.3945
> Average IOPS:           6268
> Stddev IOPS:            474.683
> Max IOPS:               6861
> Min IOPS:               5477
> Average Latency(s):     0.0102022
> Stddev Latency(s):      0.00170505
> Max latency(s):         0.0365743
> Min latency(s):         0.00641319
> Cleaning up (deleting benchmark objects)
> Removed 62727 objects
> Clean up completed and total clean up time :8.23223
> 
> 
> 
> ### SSD pool benchmark
> 
> (venv-openstack) root@os-ctrl1:~# rados -p test-ssd -t 64 -b 4096 bench 10
> write
> hints = 1
> Maintaining 64 concurrent writes of 4096 bytes to objects of size 4096 for
> up to 10 seconds or 0 objects
> Object prefix: benchmark_data_os-ctrl1_1933383
>  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg
> lat(s)
>    0       0         0         0         0         0           -
> 0
>    1      63     43839     43776   170.972       171 0.000991462
> 0.00145833
>    2      64     92198     92134   179.921   188.898  0.00211419
> 0.001387
>    3      64    141917    141853   184.675   194.215  0.00106326
> 0.00135174
>    4      63    193151    193088   188.534   200.137  0.00179379
> 0.00132423
>    5      63    243104    243041   189.847   195.129 0.000831263
> 0.00131512
>    6      63    291045    290982   189.413    187.27  0.00120208
> 0.00131807
>    7      64    341295    341231   190.391   196.285  0.00102127
> 0.00131137
>    8      63    393336    393273   191.999   203.289 0.000958149
> 0.00130041
>    9      63    442459    442396   191.983   191.887  0.00123453
> 0.00130053
> Total time run:         10.0008
> Total writes made:      488729
> Write size:             4096
> Object size:            4096
> Bandwidth (MB/sec):     190.894
> Stddev Bandwidth:       9.35224
> Max bandwidth (MB/sec): 203.289
> Min bandwidth (MB/sec): 171
> Average IOPS:           48868
> Stddev IOPS:            2394.17
> Max IOPS:               52042
> Min IOPS:               43776
> Average Latency(s):     0.00130796
> Stddev Latency(s):      0.000604629
> Max latency(s):         0.0268462
> Min latency(s):         0.000628738
> Cleaning up (deleting benchmark objects)
> Removed 488729 objects
> Clean up completed and total clean up time :8.84114
> 
> 
> 
> 
> 
> 
> 
> 
> On Wed, Aug 23, 2023 at 1:25 PM Adam King <adking@xxxxxxxxxx> wrote:
> 
>> this should be possible by specifying a "data_devices" and "db_devices"
>> fields in the OSD spec file each with different filters. There's some
>> examples in the docs
>> https://docs.ceph.com/en/latest/cephadm/services/osd/#the-simple-case that
>> show roughly how that's done, and some other sections (
>> https://docs.ceph.com/en/latest/cephadm/services/osd/#filters) that go
>> more in depth on the different filtering options available so you can try
>> and find one that works for your disks. You can check the output of "ceph
>> orch device ls --format json | jq" to see things like what cephadm
>> considers the model, size etc. for the devices to be for use in the
>> filtering.
>> 
>> On Wed, Aug 23, 2023 at 1:13 PM Satish Patel <satish.txt@xxxxxxxxx> wrote:
>> 
>>> Folks,
>>> 
>>> I have 3 nodes with each having 1x NvME (1TB) and 3x 2.9TB SSD. Trying to
>>> build ceph storage using cephadm on Ubuntu 22.04 distro.
>>> 
>>> If I want to use NvME for Journaling (WAL/DB) for my SSD based OSDs then
>>> how does cephadm handle it?
>>> 
>>> Trying to find a document where I can tell cephadm to deploy wal/db on
>>> nvme
>>> so it can speed up write optimization. Do I need to create or cephadm will
>>> create each partition for the number of OSD?
>>> 
>>> Help me to understand how it works and is it worth doing?
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>> 
>>> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux