Re: struggling to achieve high bandwidth on Ceph dev cluster - HELP

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



You have to tell a bit about your cluster setup, like nr of osd's, 3x replication on your testing pool? 

Eg. this[1] was my test on a cluster with only 1gbit ethernet, 3x repl hdd pool. This[2] with 10gbit and more osd's added

[2]
[root@c01 ~]# rados bench -p rbd 10 write
hints = 1
Maintaining 16 concurrent writes of 4194304 bytes to objects of size 4194304 for up to 10 seconds or 0 objects
Object prefix: benchmark_data_c01_3576497
  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg lat(s)
    0       0         0         0         0         0           -           0
    1      16        41        25   99.9948       100    0.198773     0.41148
    2      16       101        85   169.984       240    0.203578    0.347027
    3      16       172       156   207.979       284   0.0863202    0.296866
    4      16       245       229   228.975       292    0.139681    0.268933
    5      16       322       306   244.772       308    0.107296    0.257353
    6      16       385       369    245.97       252    0.601879    0.250782
    7      16       460       444   253.684       300    0.154803    0.247178
    8      16       541       525   262.467       324    0.274302    0.241951
    9      16       604       588     261.3       252     0.11929    0.238717
   10      16       672       656   262.367       272    0.134654    0.241424
Total time run:         10.1504
Total writes made:      673
Write size:             4194304
Object size:            4194304
Bandwidth (MB/sec):     265.212
Stddev Bandwidth:       63.0823
Max bandwidth (MB/sec): 324
Min bandwidth (MB/sec): 100
Average IOPS:           66
Stddev IOPS:            15.7706
Max IOPS:               81
Min IOPS:               25
Average Latency(s):     0.241012
Stddev Latency(s):      0.154282
Max latency(s):         1.05851
Min latency(s):         0.0702826
Cleaning up (deleting benchmark objects)
Removed 673 objects
Clean up completed and total clean up time :1.26346

[1]
[@]# rados bench -p rbd 10 write --no-cleanup
hints = 1
Maintaining 16 concurrent writes of 4194304 bytes to objects of size 4194304 for up to 10 seconds or 0 objects
Object prefix: benchmark_data_c01_18283
  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg lat(s)
    0       0         0         0         0         0           -           0
    1      16        27        11   43.9884        44    0.554119    0.624979
    2      16        47        31   61.9841        80     1.04112    0.793553
    3      16        57        41    54.654        40     1.33104    0.876273
    4      16        75        59   58.9869        72    0.840098     0.97091
    5      16        97        81   64.7864        88     1.02915    0.922043
    6      16       105        89   59.3207        32      1.2471    0.915408
    7      16       129       113   64.5582        96    0.616579    0.947882
    8      16       145       129   64.4866        64     1.09397    0.921441
    9      16       163       147   65.3201        72    0.885566    0.906388
   10      16       166       150   59.9881        12     1.22834    0.909591
   11      13       167       154   55.9889        16     2.30029    0.942798
Total time run:         11.141939
Total writes made:      167
Write size:             4194304
Object size:            4194304
Bandwidth (MB/sec):     59.9537
Stddev Bandwidth:       28.7889
Max bandwidth (MB/sec): 96
Min bandwidth (MB/sec): 12
Average IOPS:           14
Stddev IOPS:            7
Max IOPS:               24
Min IOPS:               3
Average Latency(s):     1.06157
Stddev Latency(s):      0.615773
Max latency(s):         3.23088
Min latency(s):         0.171585



> -----Original Message-----
> Sent: 10 February 2021 10:14
> To: Marc <Marc@xxxxxxxxxxxxxxxxx>
> Cc: ceph-users <ceph-users@xxxxxxx>
> Subject:  Re: struggling to achieve high bandwidth on Ceph
> dev cluster - HELP
> 
> thanks for the reply.
> 
> Yes, 4MB is the default. I have tried it. For example below (posted) is
> for
> 4MB (default) ran for 600 seconds. The seq read and rand read gives me a
> good bandwidth (not posted here). But with write its still very less.
> And I
> am particularly interested in block sizes. And rados bench tool has
> block
> size option which I have been using.
> 
> Total time run:         601.106
> Total writes made:      2966
> Write size:             4194304
> Object size:            4194304
> Bandwidth (MB/sec):     19.7369
> Stddev Bandwidth:       14.8408
> Max bandwidth (MB/sec): 64
> Min bandwidth (MB/sec): 0
> Average IOPS:           4
> Stddev IOPS:            3.67408
> Max IOPS:               16
> Min IOPS:               0
> Average Latency(s):     3.24064
> Stddev Latency(s):      2.75111
> Max latency(s):         42.4551
> Min latency(s):         0.167701
> 
> On Wed, Feb 10, 2021 at 9:46 AM Marc <Marc@xxxxxxxxxxxxxxxxx> wrote:
> 
> >
> > try 4MB that is the default not?
> >
> >
> >
> > > -----Original Message-----
> > > Sent: 10 February 2021 09:30
> > > To: ceph-users <ceph-users@xxxxxxx>; dev <dev@xxxxxxx>; ceph-
> qa@xxxxxxx
> > > Subject:  struggling to achieve high bandwidth on Ceph
> dev
> > > cluster - HELP
> > >
> > > Hi,
> > >
> > > Hello I am using rados bench tool. Currently I am using this tool
> on
> > > the
> > > development cluster after running vstart.sh script. It is working
> fine
> > > and
> > > I am interested in benchmarking the cluster. However I am struggling
> to
> > > achieve a good bandwidth i.e. bandwidth (MB/sec).  My target
> throughput
> > > is
> > > at least 50 MB/sec and more. But mostly I am achieving is around 15-
> 20
> > > MB/sec. So, very poor.
> > >
> > > I am quite sure I am missing something. Either I have to change my
> > > cluster
> > > through vstart.sh script or I am not fully utilizing the rados bench
> > > tool.
> > > Or may be both. i.e. not the right cluster and also not using the
> rados
> > > bench tool correctly.
> > >
> > > Some of the shell examples I have been using to build the cluster
> are
> > > bellow:
> > > MDS=0 RGW=1 ../src/vstart.sh -d -l -n --bluestore
> > > MDS=0 RGW=1 MON=1 OSD=4../src/vstart.sh -d -l -n --bluestore
> > >
> > > While using rados bench tool I have been trying with different block
> > > sizes
> > > 4K, 8K, 16K, 32K, 64K, 128K, 256K, 512K. And I have also been
> changing
> > > the
> > > -t parameter in the shell to increase concurrent IOs.
> > >
> >
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux