Hello, i don't understand this: Why is a rados bench with an object size of 1MB slower as a bench with object size 4MB? I attached the config in hope this helps. My version of ceph is 0.72.2 under the Debian stable distribution. The public and cluster interfaces are on a 10 GB ethernet network with jumbo frames (mtu 9000) activated. rados -p rbd bench 30 write -b 1048576 Maintaining 16 concurrent writes of 1048576 bytes for up to 30 seconds or 0 objects Object prefix: benchmark_data_ceph-mh-3_1364 sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 0 16 16 0 0 0 - 0 1 16 320 304 302.171 304 0.033876 0.0511531 2 15 613 598 298.064 294 0.027616 0.0531646 3 16 899 883 293.695 285 0.049066 0.0538804 4 15 1235 1220 304.493 337 0.025284 0.0523638 5 16 1550 1534 306.384 314 0.039108 0.0519398 6 16 1864 1848 307.646 314 0.020707 0.0518715 7 15 2156 2141 305.543 293 0.055784 0.0520902 8 16 2439 2423 302.598 282 0.098889 0.0524547 9 16 2721 2705 300.306 282 0.107799 0.0527373 10 16 2944 2928 292.576 223 0.145834 0.0543772 11 16 3233 3217 292.246 289 0.038786 0.054565 12 16 3477 3461 288.225 244 0.03195 0.0553085 13 16 3755 3739 287.436 278 0.033953 0.0555768 14 15 4043 4028 287.543 289 0.046289 0.0555471 15 16 4332 4316 287.571 288 0.04913 0.0555415 16 15 4638 4623 288.783 307 0.030713 0.0553398 17 15 4940 4925 289.539 302 0.041373 0.0551877 18 16 5203 5187 288.007 262 0.0382 0.0554286 19 15 5446 5431 285.685 244 0.036382 0.0559452 2013-12-28 17:09:49.764178min lat: 0.012016 max lat: 0.204488 avg lat: 0.0562318 sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 20 16 5690 5674 283.55 243 0.085217 0.0562318 21 15 5984 5969 284.093 295 0.109059 0.0560807 22 16 6249 6233 283.177 264 0.097483 0.056284 23 16 6506 6490 282.037 257 0.020717 0.0566822 24 16 6806 6790 282.783 300 0.02269 0.0564651 25 15 7017 7002 279.947 212 0.078625 0.0570445 26 16 7321 7305 280.832 303 0.094109 0.0569165 27 16 7631 7615 281.909 310 0.027354 0.0567099 28 15 7925 7910 282.373 295 0.038324 0.0566173 29 16 8223 8207 282.876 297 0.024542 0.0564482 30 16 8488 8472 282.279 265 0.112417 0.0565497 Total time run: 30.073757 Total writes made: 8488 Write size: 1048576 Bandwidth (MB/sec): 282.239 Stddev Bandwidth: 57.942 Max bandwidth (MB/sec): 337 Min bandwidth (MB/sec): 0 Average Latency: 0.0566741 Stddev Latency: 0.029957 Max latency: 0.311011 Min latency: 0.012016 rados -p rbd bench 30 write -b 4194304 Maintaining 16 concurrent writes of 4194304 bytes for up to 30 seconds or 0 objects Object prefix: benchmark_data_ceph-mh-3_1398 sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 0 16 16 0 0 0 - 0 1 15 111 96 383.507 384 0.220555 0.149972 2 16 223 207 413.685 444 0.150163 0.148653 3 16 324 308 410.435 404 0.06267 0.150363 4 15 414 399 398.818 364 0.258535 0.156035 5 16 513 497 397.442 392 0.165043 0.157399 6 15 612 597 397.857 400 0.227077 0.158478 7 16 689 673 384.445 304 0.2731 0.163187 8 16 750 734 366.888 244 0.13902 0.167857 9 16 847 831 369.225 388 0.104178 0.1715 10 16 962 946 378.296 460 0.167365 0.167801 11 16 1055 1039 377.719 372 0.123195 0.168256 12 16 1124 1108 369.241 276 0.244183 0.171926 13 16 1215 1199 368.833 364 0.095592 0.171366 14 16 1293 1277 364.771 312 0.160458 0.174529 15 16 1392 1376 366.85 396 0.133148 0.173578 16 16 1422 1406 351.422 120 0.744613 0.175078 17 16 1456 1440 338.75 136 0.188058 0.187931 18 16 1573 1557 345.925 468 0.138695 0.184426 19 16 1601 1585 333.614 112 0.145203 0.190529 2013-12-28 17:12:08.714308min lat: 0.045043 max lat: 1.57187 avg lat: 0.191024 sec Cur ops started finished avg MB/s cur MB/s last lat avg lat 20 16 1636 1620 323.933 140 0.120542 0.191024 21 16 1745 1729 329.266 436 0.126429 0.193505 22 16 1848 1832 333.023 412 0.083329 0.191634 23 16 1950 1934 336.28 408 0.068441 0.189515 24 15 2044 2029 338.099 380 0.222631 0.18854 25 16 2143 2127 340.253 392 0.152058 0.187696 26 16 2224 2208 339.626 324 0.168216 0.187752 27 16 2303 2287 338.748 316 0.203797 0.188278 28 16 2403 2387 340.933 400 0.101494 0.187031 29 16 2496 2480 341.979 372 0.132962 0.186758 30 16 2581 2565 341.912 340 0.204083 0.186606 Total time run: 30.178617 Total writes made: 2582 Write size: 4194304 Bandwidth (MB/sec): 342.229 Stddev Bandwidth: 115.585 Max bandwidth (MB/sec): 468 Min bandwidth (MB/sec): 0 Average Latency: 0.186847 Stddev Latency: 0.139004 Max latency: 1.57187 Min latency: 0.044427 Am 26.12.2013 17:32, schrieb Markus Martinet: > Hello, > > i have 2 OSD's and 3 MON's. Every OSD is on a 2,5TB LVM/EXT4 storage. > Why is the access to the rbd device so slow and what is meaning with > "Stddev Bandwidth" in rados bench? See the below statistics: > > #Create a 1GB file on local Storage and test a local OSD for bandwidth > cd /ceph/; dd if=/dev/zero of=test.img bs=1GB count=1 oflag=direct > 1+0 Datensätze ein > 1+0 Datensätze aus > 1000000000 Bytes (1,0 GB) kopiert, 1,8567 s, 539 MB/s > > > ceph tell osd.0 bench > { "bytes_written": 1073741824, > "blocksize": 4194304, > "bytes_per_sec": "559979881.000000"} > > > #Create a pool and test the bandwidth > rados -p vmfs bench 30 write > Maintaining 16 concurrent writes of 4194304 bytes for up to 30 seconds > or 0 objects > Object prefix: benchmark_data_ceph-mh-3_9777 > sec Cur ops started finished avg MB/s cur MB/s last lat avg lat > 0 0 0 0 0 0 - 0 > 1 16 94 78 311.879 312 0.206434 0.182676 > 2 16 191 175 349.897 388 0.177737 0.175261 > 3 16 282 266 354.578 364 0.244247 0.173725 > 4 15 372 357 356.918 364 0.167184 0.175167 > 5 16 457 441 352.72 336 0.178794 0.177867 > 6 16 553 537 357.923 384 0.244694 0.175324 > 7 16 645 629 359.351 368 0.193504 0.175503 > 8 16 725 709 354.424 320 0.235158 0.177618 > 9 15 810 795 353.244 344 0.166452 0.179567 > 10 16 888 872 348.715 308 0.15287 0.181171 > 11 16 975 959 348.64 348 0.114494 0.181629 > 12 16 1066 1050 349.864 364 0.233927 0.181363 > 13 15 1136 1121 344.795 284 0.128635 0.184239 > 14 16 1231 1215 347.019 376 0.192001 0.182952 > 15 16 1313 1297 345.747 328 0.200144 0.183385 > 16 16 1334 1318 329.389 84 0.146472 0.183347 > 17 16 1416 1400 329.303 328 0.14064 0.193126 > 18 16 1500 1484 329.67 336 0.145509 0.193292 > 19 15 1594 1579 332.315 380 0.178459 0.191693 > 2013-12-26 17:15:24.083583min lat: 0.070546 max lat: 1.17479 avg lat: > 0.191293 > sec Cur ops started finished avg MB/s cur MB/s last lat avg lat > 20 16 1665 1649 329.697 280 0.147855 0.191293 > 21 16 1701 1685 320.854 144 0.169078 0.198584 > 22 16 1750 1734 315.178 196 0.326568 0.201798 > 23 16 1805 1789 311.038 220 0.280829 0.203922 > 24 16 1882 1866 310.903 308 0.249915 0.204755 > 25 15 1966 1951 312.064 340 0.178078 0.204584 > 26 16 2030 2014 309.731 252 0.181972 0.205929 > 27 16 2085 2069 306.406 220 0.423718 0.207918 > 28 16 2179 2163 308.888 376 0.14442 0.206487 > 29 16 2252 2236 308.302 292 0.166282 0.206932 > 30 16 2325 2309 307.756 292 0.271987 0.207166 > Total time run: 30.272445 > Total writes made: 2325 > Write size: 4194304 > Bandwidth (MB/sec): 307.210 > > Stddev Bandwidth: 90.9029 > Max bandwidth (MB/sec): 388 > Min bandwidth (MB/sec): 0 > Average Latency: 0.208258 > Stddev Latency: 0.122409 > Max latency: 1.17479 > Min latency: 0.070546 > > #Test the bandwidth on a rbd device > rbd -p vmfs create test --size 1024 > rbd -p vmfs map test > dd if=/dev/rbd1 of=test.img bs=1GB count=1 oflag=direct > 1+0 Datensätze ein > 1+0 Datensätze aus > 1000000000 Bytes (1,0 GB) kopiert, 12,4647 s, 80,2 MB/s > > > Thanks, Markus > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
[global] public network = 192.168.0.3/24 cluster network = 192.168.1.3/24 auth cluster required = cephx auth service required = cephx auth client required = cephx cephx require signatures = false cephx sign messages = false keyring = /etc/ceph/$name.keyring osd pool default pg num = 100 osd pool default pgp num = 100 osd pool default size = 2 osd pool default min size = 1 [mon] mon data = /ceph/mon.$id [mds] [osd] osd data = /ceph/osd.$id # osd journal = /ceph_journal/osd.$id.journal osd journal = /ceph/osd.$id.journal osd journal size = 25600 osd pool default flag hashpspool = true osd op threads = 4 filestore xattr use omap = true filestore max sync interval = 10 [mon.0] host = ceph-mh-1 mon addr = 192.168.0.3:6789 [mon.1] host = ceph-mh-2 mon addr = 192.168.0.4:6789 [mon.2] host = ceph-mh-3 mon addr = 192.168.0.5:6789 [osd.0] host = ceph-mh-1 [osd.1] host = ceph-mh-2 [mds.0] host = ceph-mh-1 [mds.1] host = ceph-mh-2 [client] rbd cache = true
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com