Re: Large latency for single thread

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Is this not just inherent to SDS? And wait for the new osd code, I think they are working on it.

https://yourcmc.ru/wiki/Ceph_performance


> 
> m-seqwr-004k-001q-001j: (groupid=0, jobs=1): err= 0: pid=46: Wed Dec 15
> 14:05:32 2021
>    write: IOPS=794, BW=3177KiB/s (3254kB/s)(559MiB/180002msec); 0 zone
> resets
>      slat (usec): min=4, max=123, avg=22.30, stdev= 9.18
>      clat (usec): min=630, max=16977, avg=1232.89, stdev=354.67
>       lat (usec): min=639, max=17009, avg=1255.19, stdev=358.99
>      clat percentiles (usec):
>       |  1.00th=[  709],  5.00th=[  775], 10.00th=[  824],
> 20.00th=[  906],
>       | 30.00th=[ 1074], 40.00th=[ 1172], 50.00th=[ 1237], 60.00th=[
> 1303],
>       | 70.00th=[ 1369], 80.00th=[ 1450], 90.00th=[ 1565], 95.00th=[
> 1663],
>       | 99.00th=[ 2606], 99.50th=[ 3261], 99.90th=[ 3785], 99.95th=[
> 3949],
>       | 99.99th=[ 6718]
>     bw (  KiB/s): min= 1928, max= 5048, per=100.00%, avg=3179.54,
> stdev=588.79, samples=360
>     iops        : min=  482, max= 1262, avg=794.76, stdev=147.20,
> samples=360
>    lat (usec)   : 750=2.98%, 1000=22.41%
>    lat (msec)   : 2=73.38%, 4=1.18%, 10=0.04%, 20=0.01%
>    cpu          : usr=2.69%, sys=1.78%, ctx=145218, majf=0, minf=2
>    IO depths    : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%,
>  >=64=0.0%
>       submit    : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
>  >=64=0.0%
>       complete  : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%,
>  >=64=0.0%
>       issued rwts: total=0,142985,0,0 short=0,0,0,0 dropped=0,0,0,0
>       latency   : target=0, window=0, percentile=100.00%, depth=1
> 
> 
> Parts of the OSD' perf status:
> 
>       "state_io_done_lat": {
>              "avgcount": 151295,
>              "sum": 0.336297058,
>              "avgtime": 0.000002222
>          },
>          "state_kv_queued_lat": {
>              "avgcount": 151295,
>              "sum": 18.812333051,
>              "avgtime": 0.000124342
>          },
>          "state_kv_commiting_lat": {
>              "avgcount": 151295,
>              "sum": 64.555436175,
>              "avgtime": 0.000426685
>          },
>          "state_kv_done_lat": {
>              "avgcount": 151295,
>              "sum": 0.130403628,
>              "avgtime": 0.000000861
>          },
>          "state_deferred_queued_lat": {
>              "avgcount": 148,
>              "sum": 215.726286547,
>              "avgtime": 1.457610044
>          },
> 
> ... ...
> 
>          "op_w_latency": {
>              "avgcount": 151133,
>              "sum": 130.134246667,
>              "avgtime": 0.000861057
>          },
>          "op_w_process_latency": {
>              "avgcount": 151133,
>              "sum": 125.301196872,
>              "avgtime": 0.000829079
>          },
>          "op_w_prepare_latency": {
>              "avgcount": 151133,
>              "sum": 29.892687947,
>              "avgtime": 0.000197790
>          },
> 
> Is it reasonable for the benchmark test case?  And how to improve it?
> It's really NOT friendly for single thread.
> 
> 
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux