Ceph Tuning + KV backend

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

Excuse my ignorance, I have just joined this list and started using Ceph (which looks very cool). On AWS I have set up a 5-way Ceph cluster (4 vCPUs, 32G RAM, dedicated SSDs for system, osd and journal) with the Object Gateway. For the purpose of simplicity of the test all the nodes are identical and each node contains osd, mon and the radosgw.

I have run parallel inserts from all 5 nodes, I can insert about 10-12000 objects per minute. The insert rate is relatively constant regardless of whether I run 1 insert process per node or 5, i.e. a total of 5 or 25.

These are just numbers, of course, and not meaningful without more context. But looking at the nodes I think the cluster could run faster - the CPUs are not doing much, there isn't much I/O wait - only about 50% utilisation and only on the SSDs storing the journals on two of the nodes (I've set the replication to 2), the other file systems are almost idle. The network is far from maxed out and the processes are not using much memory. I've tried increasing osd_op_threads to 5 or 10 but that didn't make much difference.

The co-location of all the daemons on all the nodes may not be ideal, but since there isn't much resource use or contention I don't think that's the problem.

So two questions:

1) Are there any good resources on tuning Ceph? There's quite a few posts out there testing and timing specific setups with RAID controller X and 12 disks of brand Y etc. but I'm more looking for general tuning guidelines - explaining the big picture.

2) What's the status of the keyvalue backend? The documentation on http://ceph.com/docs/master/rados/configuration/keyvaluestore-config-ref/ looks nice but I found it difficult to work out how to switch to the keyvalue backend, the Internet suggests "osd objectstore = keyvaluestore-dev", but that didn't seem to work so I checked out the source code and it looks like "osd objectstore = keyvaluestore" does it. However, it results in nasty things in the log file ("*** experimental feature 'keyvaluestore' is not enabled *** This feature is marked as experimental ...") so perhaps it's too early to use the KV backend for production use?

Thanks & regards,
Jakob
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux