osd store. Notice that your tested throughput is slightly more than half your expected maximum...
Hi all,
I run 3 nodes connected with a 10Gbit network, each running 2 OSDs.
Disks are 4TB Seagate Constellation ST4000NM0033-9ZM (xfs, journal on same disk).
# ceph tell osd.0 bench
{ "bytes_written": 1073741824,
"blocksize": 4194304,
"bytes_per_sec": "56494242.000000"}
So a single OSD can write up to OSD_BW=55MB/s
I have 6 OSDs, so I would expect an overall write bandwidth of:
6*OSD_BW = 6*55 = 330MB/s
To verify that, I created a pool with size=1, and run:
# rados -p testpool bench 300 write --no-cleanup
…
Total time run: 300.905982
Total writes made: 13487
Write size: 4194304
Bandwidth (MB/sec): 179.285
Stddev Bandwidth: 106.954
Max bandwidth (MB/sec): 400
Min bandwidth (MB/sec): 0
Average Latency: 0.356925
Stddev Latency: 0.696282
Max latency: 8.5891
Min latency: 0.025648
This is much lower than the expected bandwidth (179 < 330).
Is this normal? If so, what is the reason for that?
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
--
______________________________
Dino Yancey
2GNT.com Admin
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com