Re: Ceph Tuning + KV backend

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Sep 9, 2015 at 3:00 AM, Niels Jakob Darger <jakob@xxxxxxxxxx> wrote:
> Hello,
>
> Excuse my ignorance, I have just joined this list and started using Ceph
> (which looks very cool). On AWS I have set up a 5-way Ceph cluster (4 vCPUs,
> 32G RAM, dedicated SSDs for system, osd and journal) with the Object
> Gateway. For the purpose of simplicity of the test all the nodes are
> identical and each node contains osd, mon and the radosgw.
>
> I have run parallel inserts from all 5 nodes, I can insert about 10-12000
> objects per minute. The insert rate is relatively constant regardless of
> whether I run 1 insert process per node or 5, i.e. a total of 5 or 25.
>
> These are just numbers, of course, and not meaningful without more context.
> But looking at the nodes I think the cluster could run faster - the CPUs are
> not doing much, there isn't much I/O wait - only about 50% utilisation and
> only on the SSDs storing the journals on two of the nodes (I've set the
> replication to 2), the other file systems are almost idle. The network is
> far from maxed out and the processes are not using much memory. I've tried
> increasing osd_op_threads to 5 or 10 but that didn't make much difference.
>
> The co-location of all the daemons on all the nodes may not be ideal, but
> since there isn't much resource use or contention I don't think that's the
> problem.
>
> So two questions:
>
> 1) Are there any good resources on tuning Ceph? There's quite a few posts
> out there testing and timing specific setups with RAID controller X and 12
> disks of brand Y etc. but I'm more looking for general tuning guidelines -
> explaining the big picture.
>
> 2) What's the status of the keyvalue backend? The documentation on
> http://ceph.com/docs/master/rados/configuration/keyvaluestore-config-ref/
> looks nice but I found it difficult to work out how to switch to the
> keyvalue backend, the Internet suggests "osd objectstore =
> keyvaluestore-dev", but that didn't seem to work so I checked out the source
> code and it looks like "osd objectstore = keyvaluestore" does it. However,
> it results in nasty things in the log file ("*** experimental feature
> 'keyvaluestore' is not enabled *** This feature is marked as experimental
> ...") so perhaps it's too early to use the KV backend for production use?
>

Hmm, we need to modify doc to swith to "keyvaluestore". Since this is
a experimental feature, it doesn't recommended to used for production
usage.

> Thanks & regards,
> Jakob
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



-- 
Best Regards,

Wheat
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux