Designing a cluster with ceph and benchmark (ceph vs ext4)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks Cristian. I will reflect on what you told me. There is no free 
lunch, I'll think it's worth paying the price.

--
Thiago Henrique

Em 24-05-2014 02:43, Christian Balzer escreveu:
>
> Hello,
>
> On Fri, 23 May 2014 15:41:23 -0300 Listas at Adminlinux wrote:
>
>> Hi !
>>
>> I have failover clusters for some aplications. Generally with 2 members
>> configured with Ubuntu + Drbd + Ext4. For example, my IMAP cluster works
>> fine with ~ 50k email accounts and my HTTP cluster hosts ~2k sites.
>>
> My mailbox servers are also multiple DRBD based cluster pairs.
> For performance in fully redundant storage there is isn't anything better
> (in the OSS, generic hardware section at least).
>
>> See design here: http://adminlinux.com.br/cluster_design.txt
>>
>> I would like to provide load balancing instead of just failover. So, I
>> would like to use a distributed architecture of the filesystem. As we
>> know, Ext4 isn't a distributed filesystem. So wish to use Ceph in my
>> clusters.
>>
> You will find that all cluster/distributed filesystems have severe
> performance shortcomings when compared to something like Ext4.
>
> On top of that, CephFS isn't ready for production as the MDS isn't HA.
>
> A potential middle way might be to use Ceph/RBD volumes formatted in Ext4.
> That doesn't give you shared access, but it will allow you to separate
> storage and compute nodes, so when one compute node becomes busy, mount
> that volume from a more powerful compute node instead.
>
> That all said, I can't see any way and reason to replace my mailbox DRBD
> clusters with Ceph in the foreseeable future.
> To get similar performance/reliability to DRBD I would have to spend 3-4
> times the money.
>
> Where Ceph/RBD works well is situations where you can't fit the compute
> needs into a storage node (as required with DRBD) and where you want to
> access things from multiple compute nodes, primarily for migration
> purposes.
> In short, as a shared storage for VMs.
>
>> Any suggestions for design of the cluster with Ubuntu+Ceph?
>>
>> I built a simple cluster of 2 servers to test simultaneous reading and
>> writing with Ceph. My conf:  http://adminlinux.com.br/ceph_conf.txt
>>
> Again, CephFS isn't ready for production, but other than that I know very
> little about it as I don't use it.
> However your version of Ceph is severely outdated, you really should be
> looking at something more recent to rule out you're experience long fixed
> bugs. The same goes for your entire setup and kernel.
>
> Also Ceph only starts to perform decently with many OSDs (disks) and
> the journals on SSDs instead of being on the same disk.
> Think DRBD AL metadata-internal, but with MUCH more impact.
>
> Regards,
>
> Christian
>> But in my simultaneous benchmarks found errors in reading and writing. I
>> ran "iozone -t 5 -r 4k -s 2m" simultaneously on both servers in the
>> cluster. The performance was poor and had errors like this:
>>
>> Error in file: Found ?0? Expecting ?6d6d6d6d6d6d6d6d? addr b6600000
>> Error in file: Position 1060864
>> Record # 259 Record size 4 kb
>> where b6600000 loop 0
>>
>> Performance graphs of benchmark: http://adminlinux.com.br/ceph_bench.html
>>
>> Can you help me find what I did wrong?
>>
>> Thanks !
>>
>
>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux