Re: ceph bluestore speed

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



> 
> I'm trying to set up a test lab with ceph (on proxmox).
> I've got 3 nodes, but I figured I'd start with 1 to test out speeds and
> to
> learn more about the setup of ceph. I will add the 2 extra nodes later.
> 
> One thing that was disappointing was the writing speed.
> 
> In my setup I've got 14 * 300GB SAS hdd drives. When I do a write test
> directly on one of them I get around 130-140MB/sec write speed:
> fio --filename=/dev/sdp1 -name=test -direct=1 -rw=write -bs=4M -
> iodepth=16
> Run status group 0 (all jobs):
>   WRITE: bw=132MiB/s (138MB/s), 132MiB/s-132MiB/s (138MB/s-138MB/s),
> io=153GiB (164GB), run=1187061-1187061msec

You can forget about this sequential write test. Ceph is writing random.


> When I set up ceph osd.0 trough osd.13 (hdd bluestore - DB/WAL on each
> osd), create a pool that I can use and then run a write speed:
> fio -ioengine=rbd -name=test -direct=1 -rw=write -bs=4M -iodepth=16
> -pool=ceph -rbdname=vm-118-disk-0
> I get only around 80MB/sec:
> Run status group 0 (all jobs):
>   WRITE: bw=82.4MiB/s (86.4MB/s), 82.4MiB/s-82.4MiB/s (86.4MB/s-
> 86.4MB/s),
> io=100GiB (107GB), run=1243230-1243230msec
> 
> Am I doing something wrong, or is the write speed supposed to drop when
> clustering disks like this?
> 

Probably not. You never have native disk speed, because your data is distributed over multiple disks on multiple nodes, that just takes its time.
If you do a this rados bench, you can compare with my (very) slow hdd drives.

[~]# rados bench -p rbd 10 write
hints = 1
Maintaining 16 concurrent writes of 4194304 bytes to objects of size 4194304 for up to 10 seconds or 0 objects
Object prefix: benchmark_data_c01_538834
  sec Cur ops   started  finished  avg MB/s  cur MB/s last lat(s)  avg lat(s)
    0       0         0         0         0         0           -           0
    1      16        77        61   243.933       244    0.273347    0.225817
    2      16       130       114   227.952       212    0.184422    0.242989
    3      16       182       166   221.287       208    0.132519    0.261817
    4      16       234       218   217.958       208    0.145279    0.272812
    5      16       302       286   228.757       272   0.0905078    0.265453
    6      16       370       354   235.957       272    0.260377    0.268706
    7      16       414       398   227.389       176      1.0019    0.271782
    8      16       474       458   228.961       240    0.120364    0.271026
    9      15       541       526   233.739       272    0.120498    0.266636
   10      16       595       579   231.561       212     0.11489    0.269708
Total time run:         10.2101
Total writes made:      596
Write size:             4194304
Object size:            4194304
Bandwidth (MB/sec):     233.495
Stddev Bandwidth:       33.4903
Max bandwidth (MB/sec): 272
Min bandwidth (MB/sec): 176
Average IOPS:           58
Stddev IOPS:            8.37257
Max IOPS:               68
Min IOPS:               44
Average Latency(s):     0.273409
Stddev Latency(s):      0.195287
Max latency(s):         1.23514
Min latency(s):         0.0680838
Cleaning up (deleting benchmark objects)
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux