Re: Ceph instead of RAID

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Aug 13, 2013 at 10:41:53AM -0500, Mark Nelson wrote:

Hi Mark,

> On 08/13/2013 02:56 AM, Dmitry Postrigan wrote:
> >>>I am currently installing some backup servers with 6x3TB drives in them. I played with RAID-10 but I was not
> >>>impressed at all with how it performs during a recovery.
> >>>
> >>>Anyway, I thought what if instead of RAID-10 I use ceph? All 6 disks will be local, so I could simply create
> >>>6 local OSDs + a monitor, right? Is there anything I need to watch out for in such configuration?
> >
> >>You can do that. Although it's nice to play with and everything, I
> >>wouldn't recommend doing it. It will give you more pain than pleasure.
> >
> >Any specific reason? I just got it up and running, an after simulating some failures, I like it much better than
> >mdraid. Again, this only applies to large arrays (6x3TB in my case). I would not use ceph to replace a RAID-1
> >array of course, but it looks like a good idea to replace a large RAID10 array with a local ceph installation.
> >
> >The only thing I do not enjoy about ceph is performance. Probably need to do more tweaking, but so far numbers
> >are not very impressive. I have two exactly same servers running same OS, kernel, etc. Each server has 6x 3TB
> >drives (same model and firmware #).
> >
> >Server 1 runs ceph (2 replicas)
> >Server 2 runs mdraid (raid-10)
> >
> >I ran some very basic benchmarks on both servers:
> >
> >dd if=/dev/zero of=/storage/test.bin bs=1M count=100000
> >Ceph: 113 MB/s
> >mdraid: 467 MB/s
> >
> >
> >dd if=/storage/test.bin of=/dev/null bs=1M
> >Ceph: 114 MB/s
> >mdraid: 550 MB/s
> >
> >
> >As you can see, mdraid is by far faster than ceph. It could be "by design", or perhaps I am not doing it
> >right. Even despite such difference in speed, I would still go with ceph because *I think* it is more reliable.
> 
> couple of things:
> 
> 1) Ceph is doing full data journal writes so is going to eat (at
> least) half of your write performance right there.
> 
> 2) Ceph tends to like lots of concurrency.  You'll probably see
> higher numbers with multiple dd reads/writes going at once.
> 
> 3) Ceph is a lot more complex than something like mdraid.  It gives
> you a lot more power and flexibility but the cost is greater
> complexity. There are probably things you can tune to get your
> numbers up, but it could take some work.
> 
> Having said all of this, my primary test box is a single server and
> I can get 90MB/s+ per drive out of Ceph (with 24 drives!), but if I

Could you share the configurations and parameters you have modified, or
where I could find the associate documents? 

> was building a production box and never planned to expand to
> multiple servers, I'd certainly be looking into zfs or btrfs RAID.
> 
> Mark
> 
> >
> >Dmitry
> >
> >_______________________________________________
> >ceph-users mailing list
> >ceph-users@xxxxxxxxxxxxxx
> >http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
> 
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

-- 
Best regards,
Guangliang
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux