Bob, Those numbers would seem to indicate some other problem .... One of the biggest culprits of that poor performance is often related to network issues. In the last few months, there have been several reported issues of performance, that have turned out to be network. Not all, but most. You're best bet is to check each host interface statistics for errors. make sure you have a match on the MTU size (jumbo frames settings on the host and on your switches). Check your switches for network errors. Try extended size ping checks between nodes, insure you set the packet size close to your max MTU size and check that you're getting good performance from *all nodes* to every other node. Last, try a network performance test to each of the OSD nodes and see if one of them is acting up. If you are backing your journal on SSD - you DEFINITELY should be getting vastly better performance than that. I have a cluster with 6 OSD nodes w/ 10x 4TB OSDs - using 2 7200 rpm disks as the journatl (12 disks total). NO SSDs in that configuration. I can push the cluster to about 650 MByte/sec via network RBD 'dd' tests, and get about 2500 IOPS. NOTE - this is an all spinning HDD cluster w/ 7200 rpm disks! ~~shane On 8/4/15, 2:36 PM, "ceph-users on behalf of Bob Ababurko" <ceph-users-bounces@xxxxxxxxxxxxxx on behalf of bob@xxxxxxxxxxxx> wrote:
|
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com