Hi guys,
I am currently testing the performance of RBD using a cache pool and a 4/2 erasure profile pool.
I have two SSD cache servers (2 SSDs for journals, 7 SSDs for data) with 2x10Gbit bonded each and a six OSD nodes with a 10Gbit public and 10Gbit cluster network for the erasure pool (10x3TB without separate journal). This is all on Jewel.
What I would like to know is if the performance I'm seeing is to be expected and if there is some way to test this in a more qualifiable way.
Everything works as expected if the files are present on the cache pool, however when things need to be retrieved from the cache pool I see performance degradation. I'm trying to simulate real usage as much as possible and trying to retrieve files from the RBD volume over FTP from a client server. What I'm seeing is that the FTP transfer will stall for seconds at a time and then get some more data which results in an average speed of 200KB/s. From the cache this is closer to 10MB/s. Is this the expected behaviour from a erasure coded tier with cache in front?
Right now I'm unsure how to scientifically test the performance retrieving files when there is a cache miss. If somebody could point me towards a better way of doing that I would appreciate the help.
An other thing is that I'm seeing a lot of messages popping up in dmesg on my client server on which the RBD volumes are mounted. (IPs removed)
[685881.477383] libceph: osd50 :6800 socket closed (con state OPEN)
[685895.597733] libceph: osd54 :6808 socket closed (con state OPEN)
[685895.663971] libceph: osd54 :6808 socket closed (con state OPEN)
[685895.710424] libceph: osd54 :6808 socket closed (con state OPEN)
[685895.749417] libceph: osd54 :6808 socket closed (con state OPEN)
[685896.517778] libceph: osd54 :6808 socket closed (con state OPEN)
[685906.690445] libceph: osd74 :6824 socket closed (con state OPEN)
Is this a symptom of something?
Thanks in advance,
Peter
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com