Ok! Well, I'll take these tests and write about the results. btw, disks are the same, as some may be faster than others? 2012/11/4 Gregory Farnum <greg@xxxxxxxxxxx>: > That's only nine — where are the other three? If you have three slow > disks that could definitely cause the troubles you're seeing. > > Also, what Mark said about sync versus syncfs. > > On Sun, Nov 4, 2012 at 1:26 PM, Aleksey Samarin <nrg3tik@xxxxxxxxx> wrote: >> It`s ok! >> >> Output: >> >> 2012-11-04 16:19:23.195891 osd.0 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 11.441035 sec at 91650 KB/sec >> 2012-11-04 16:19:24.981631 osd.1 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 13.225048 sec at 79287 KB/sec >> 2012-11-04 16:19:25.672896 osd.2 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 13.917157 sec at 75344 KB/sec >> 2012-11-04 16:19:28.058517 osd.21 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 16.453375 sec at 63730 KB/sec >> 2012-11-04 16:19:28.715552 osd.22 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 17.108887 sec at 61288 KB/sec >> 2012-11-04 16:19:23.440054 osd.23 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 11.834639 sec at 88602 KB/sec >> 2012-11-04 16:19:24.023650 osd.24 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 12.418276 sec at 84438 KB/sec >> 2012-11-04 16:19:24.617514 osd.25 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 13.011955 sec at 80585 KB/sec >> 2012-11-04 16:19:25.148613 osd.26 [INF] bench: wrote 1024 MB in blocks >> of 4096 KB in 13.541710 sec at 77433 KB/sec >> >> All the best. >> >> 2012/11/4 Gregory Farnum <greg@xxxxxxxxxxx>: >>> [Sorry for the blank email; I missed!] >>> On Sun, Nov 4, 2012 at 1:04 PM, Aleksey Samarin <nrg3tik@xxxxxxxxx> wrote: >>>> Hi! >>>> This command? ceph tell osd \* bench >>>> Output: tell target 'osd' not a valid entity name >>> >>> I guess it's "ceph osd tell \* bench". Try that one. :) >>> >>>> Well, i did pool by command ceph osd pool create bench2 120 >>>> This output of rados -p bench2 bench 30 write --no-cleanup >>>> >>>> rados -p bench2 bench 30 write --no-cleanup >>>> >>>> Maintaining 16 concurrent writes of 4194304 bytes for at least 30 seconds. >>>> Object prefix: benchmark_data_host01_5827 >>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat >>>> 0 0 0 0 0 0 - 0 >>>> 1 16 29 13 51.9885 52 0.489268 0.186749 >>>> 2 16 52 36 71.9866 92 1.87226 0.711888 >>>> 3 16 57 41 54.657 20 0.089697 0.697821 >>>> 4 16 60 44 43.9923 12 1.61868 0.765361 >>>> 5 16 60 44 35.1941 0 - 0.765361 >>>> 6 16 60 44 29.3285 0 - 0.765361 >>>> 7 16 60 44 25.1388 0 - 0.765361 >>>> 8 16 61 45 22.4964 1 5.89643 0.879384 >>>> 9 16 62 46 20.4412 4 6.0234 0.991211 >>>> 10 16 62 46 18.3971 0 - 0.991211 >>>> 11 16 63 47 17.0883 2 8.79749 1.1573 >>>> 12 16 63 47 15.6643 0 - 1.1573 >>>> 13 16 63 47 14.4593 0 - 1.1573 >>>> 14 16 63 47 13.4266 0 - 1.1573 >>>> 15 16 63 47 12.5315 0 - 1.1573 >>>> 16 16 63 47 11.7483 0 - 1.1573 >>>> 17 16 63 47 11.0572 0 - 1.1573 >>>> 18 16 63 47 10.4429 0 - 1.1573 >>>> 19 16 63 47 9.89331 0 - 1.1573 >>>> 2012-11-04 15:58:15.473733min lat: 0.036475 max lat: 8.79749 avg lat: 1.1573 >>>> sec Cur ops started finished avg MB/s cur MB/s last lat avg lat >>>> 20 16 63 47 9.39865 0 - 1.1573 >>>> 21 16 63 47 8.95105 0 - 1.1573 >>>> 22 16 63 47 8.54419 0 - 1.1573 >>>> 23 16 63 47 8.17271 0 - 1.1573 >>>> 24 16 63 47 7.83218 0 - 1.1573 >>>> 25 16 63 47 7.5189 0 - 1.1573 >>>> 26 16 63 47 7.22972 0 - 1.1573 >>>> 27 16 81 65 9.62824 4.5 0.076456 4.9428 >>>> 28 16 118 102 14.5693 148 0.427273 4.34095 >>>> 29 16 119 103 14.2049 4 1.57897 4.31414 >>>> 30 16 132 116 15.4645 52 2.25424 4.01492 >>>> 31 16 133 117 15.0946 4 0.974652 3.98893 >>>> 32 16 133 117 14.6229 0 - 3.98893 >>>> Total time run: 32.575351 >>>> Total writes made: 133 >>>> Write size: 4194304 >>>> Bandwidth (MB/sec): 16.331 >>>> >>>> Stddev Bandwidth: 31.8794 >>>> Max bandwidth (MB/sec): 148 >>>> Min bandwidth (MB/sec): 0 >>>> Average Latency: 3.91583 >>>> Stddev Latency: 7.42821 >>>> Max latency: 25.24 >>>> Min latency: 0.036475 >>>> >>>> Im think problem not in pg. This output of ceph pg dump > >>>> http://pastebin.com/BqLsyMBC >>> >>> Well, that did improve it a bit; but yes, I think there's something >>> else going on. Just wanted to verify. :) >>> >>>> >>>> I have still no idea. >>>> >>>> All the best. Alex >>>> >>>> >>>> >>>> 2012/11/4 Gregory Farnum <greg@xxxxxxxxxxx>: >>>>> On Sun, Nov 4, 2012 at 10:58 AM, Aleksey Samarin <nrg3tik@xxxxxxxxx> wrote: >>>>>> Hi all >>>>>> >>>>>> Im planning use ceph for cloud storage. >>>>>> My test setup is 2 servers connected via infiniband 40Gb, 6x2Tb disks per node. >>>>>> Centos 6.2 >>>>>> Ceph 0.52 from http://ceph.com/rpms/el6/x86_64 >>>>>> This is my config http://pastebin.com/Pzxafnsm >>>>>> journal on tmpfs >>>>>> well, im create bench pool and test it: >>>>>> ceph osd pool create bench >>>>>> rados -p bench bench 30 write >>>>>> >>>>>> Total time run: 43.258228 >>>>>> Total writes made: 151 >>>>>> Write size: 4194304 >>>>>> Bandwidth (MB/sec): 13.963 >>>>>> Stddev Bandwidth: 26.307 >>>>>> Max bandwidth (MB/sec): 128 >>>>>> Min bandwidth (MB/sec): 0 >>>>>> Average Latency: 4.48605 >>>>>> Stddev Latency: 8.17709 >>>>>> Max latency: 29.7957 >>>>>> Min latency: 0.039435 >>>>>> >>>>>> when i do rados -p bench bench 30 seq >>>>>> Total time run: 20.626935 >>>>>> Total reads made: 275 >>>>>> Read size: 4194304 >>>>>> Bandwidth (MB/sec): 53.328 >>>>>> Average Latency: 1.19754 >>>>>> Max latency: 7.0215 >>>>>> Min latency: 0.011647 >>>>>> >>>>>> I tested the single drive via dd if=/dev/zero of=/mnt/hdd2/testfile >>>>>> bs=1024k count=20000 >>>>>> result: 158 MB/sec >>>>>> >>>>>> Anyone can tell me why such a weak performance? Maybe I missed something? >>>>> >>>>> Can you run "ceph tell osd \* bench" and report the results? (It'll go >>>>> to the "central log" which you can keep an eye on if you run "ceph -w" >>>>> in another terminal.) >>>>> I think you also didn't create your bench pool correctly; it probably >>>>> only has 8 PGs which is not going to perform very well with your disk >>>>> count. Try "ceph pool create bench2 120" and run the benchmark against >>>>> that pool. The extra number at the end tells it to create 120 >>>>> placement groups. >>>>> -Greg -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html