On Wed, Sep 9, 2015 at 3:00 AM, Niels Jakob Darger <jakob@xxxxxxxxxx> wrote: > Hello, > > Excuse my ignorance, I have just joined this list and started using Ceph > (which looks very cool). On AWS I have set up a 5-way Ceph cluster (4 vCPUs, > 32G RAM, dedicated SSDs for system, osd and journal) with the Object > Gateway. For the purpose of simplicity of the test all the nodes are > identical and each node contains osd, mon and the radosgw. > > I have run parallel inserts from all 5 nodes, I can insert about 10-12000 > objects per minute. The insert rate is relatively constant regardless of > whether I run 1 insert process per node or 5, i.e. a total of 5 or 25. > > These are just numbers, of course, and not meaningful without more context. > But looking at the nodes I think the cluster could run faster - the CPUs are > not doing much, there isn't much I/O wait - only about 50% utilisation and > only on the SSDs storing the journals on two of the nodes (I've set the > replication to 2), the other file systems are almost idle. The network is > far from maxed out and the processes are not using much memory. I've tried > increasing osd_op_threads to 5 or 10 but that didn't make much difference. > > The co-location of all the daemons on all the nodes may not be ideal, but > since there isn't much resource use or contention I don't think that's the > problem. > > So two questions: > > 1) Are there any good resources on tuning Ceph? There's quite a few posts > out there testing and timing specific setups with RAID controller X and 12 > disks of brand Y etc. but I'm more looking for general tuning guidelines - > explaining the big picture. > > 2) What's the status of the keyvalue backend? The documentation on > http://ceph.com/docs/master/rados/configuration/keyvaluestore-config-ref/ > looks nice but I found it difficult to work out how to switch to the > keyvalue backend, the Internet suggests "osd objectstore = > keyvaluestore-dev", but that didn't seem to work so I checked out the source > code and it looks like "osd objectstore = keyvaluestore" does it. However, > it results in nasty things in the log file ("*** experimental feature > 'keyvaluestore' is not enabled *** This feature is marked as experimental > ...") so perhaps it's too early to use the KV backend for production use? > Hmm, we need to modify doc to swith to "keyvaluestore". Since this is a experimental feature, it doesn't recommended to used for production usage. > Thanks & regards, > Jakob > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Best Regards, Wheat _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com