> -----Original Message----- > From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of > Heath Albritton > Sent: 23 May 2016 01:24 > To: ceph-users@xxxxxxxxxxxxxx > Subject: Re: NVRAM cards as OSD journals > > I'm contemplating the same thing as well. Or rather, I'm actually doing some > testing. I have a Netlist EV3 and have seen ~6GB/s read and write for any > block size larger than 16k or so, IIRC. > > Sebastien Han has a blog page with journal benchmarks, I've added the > specifics there. > > This week, I expect to test the performance with ceph on a single node. I > have two identically configured nodes, each with 8 SATA SSDs as OSDs. My > goal is to see if the nvram card improves the performance, specifically with > regard to latency. > > The Netlist folks claim significantly improved performance for 64k blocks with > ceph. Thy and reduced latency bodes well for database type use cases. A while back for a test, I ran a journal on a kernel zram device and managed to get qd=1 bs=4kb write IO's up to about 2200iops. So Write latency was under 0.5ms. Would be very interested in seeing your results with the NVRAM cards. > > -H > > On May 20, 2016, at 08:52, EP Komarla <Ep.Komarla@xxxxxxxxxxxxxxx> wrote: > Hi, > > I am contemplating using a NVRAM card for OSD journals in place of SSD > drives in our ceph cluster. > > Configuration: > • 4 Ceph servers > • Each server has 24 OSDs (each OSD is a 1TB SAS drive) > • 1 PCIe NVRAM card of 16GB capacity per ceph server > • Both Client & cluster network is 10Gbps > > As per ceph documents: > The expected throughput number should include the expected disk > throughput (i.e., sustained data transfer rate), and network throughput. For > example, a 7200 RPM disk will likely have approximately 100 MB/s. Taking > the min() of the disk and network throughput should provide a reasonable > expected throughput. Some users just start off with a 10GB journal size. For > example: > osd journal size = 10000 > Given that I have a single 16GB card per server that has to be carved among > all 24OSDs, I will have to configure each OSD journal to be much smaller > around 600MB, i.e., 16GB/24 drives. This value is much smaller than > 10GB/OSD journal that is generally used. So, I am wondering if this > configuration and journal size is valid. Is there a performance benefit of > having a journal that is this small? Also, do I have to reduce the default > “filestore maxsync interval” from 5 seconds to a smaller value say 2 seconds > to match the smaller journal size? > > Have people used NVRAM cards in the Ceph clusters as journals? What is > their experience? > > Any thoughts? > > > > Legal Disclaimer: > The information contained in this message may be privileged and > confidential. It is intended to be read only by the individual or entity to whom > it is addressed or by their designee. If the reader of this message is not the > intended recipient, you are on notice that any distribution of this message, in > any form, is strictly prohibited. If you have received this message in error, > please immediately notify the sender and delete or destroy any copy of this > message! > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com