Thanks >> 8x SSD, 200GB each 20000 iops seem pretty low,no ? for @intank: Is their a bottleneck somewhere in ceph ? I said that, because I would like to know if it's scale by adding new nodes. Does Intank have already done some random iops benchmark ? (I always see sequential throughput bench in the mailing list) ----- Mail original ----- De: "Dieter Kasper" <d.kasper@xxxxxxxxxxxx> À: "Alexandre DERUMIER" <aderumier@xxxxxxxxx> Cc: ceph-devel@xxxxxxxxxxxxxxx Envoyé: Jeudi 30 Août 2012 17:33:42 Objet: Re: RBD performance - tuning hints On Thu, Aug 30, 2012 at 05:28:02PM +0200, Alexandre DERUMIER wrote: > Thanks for the report ! > > vs your first benchmark, it's with RBD 4M or 64K ? with 4MB (see attached config info) Cheers, -Dieter > > (how much ssd by node?) 8x SSD, 200GB each > > > > ----- Mail original ----- > > De: "Dieter Kasper" <d.kasper@xxxxxxxxxxxx> > À: "Alexandre DERUMIER" <aderumier@xxxxxxxxx> > Cc: ceph-devel@xxxxxxxxxxxxxxx > Envoyé: Jeudi 30 Août 2012 16:56:34 > Objet: Re: RBD performance - tuning hints > > Hi Alexandre, > > with the 4 filestore parameter below some fio values could be increased: > filestore max sync interval = 30 > filestore min sync interval = 29 > filestore flusher = false > filestore queue max ops = 10000 > > ###### IOPS > fio_read_4k_64: 9373 > fio_read_4k_128: 9939 > fio_randwrite_8k_16: 12376 > fio_randwrite_4k_16: 13315 > fio_randwrite_512_32: 13660 > fio_randwrite_8k_32: 17318 > fio_randwrite_4k_32: 18057 > fio_randwrite_8k_64: 19693 > fio_randwrite_512_64: 20015 <<< > fio_randwrite_4k_64: 20024 <<< > fio_randwrite_8k_128: 20547 <<< > fio_randwrite_4k_128: 20839 <<< > fio_randwrite_512_128: 21417 <<< > fio_randread_8k_128: 48872 > fio_randread_4k_128: 50002 > fio_randread_512_128: 51202 > > ###### MB/s > fio_randread_2m_32: 628 > fio_read_4m_64: 630 > fio_randread_8m_32: 633 > fio_read_2m_32: 637 > fio_read_4m_16: 640 > fio_randread_4m_16: 652 > fio_write_2m_32: 660 > fio_randread_4m_32: 677 > fio_read_4m_32: 678 > (...) > fio_write_4m_64: 771 > fio_randwrite_2m_64: 789 > fio_write_8m_128: 796 > fio_write_4m_32: 802 > fio_randwrite_4m_128: 807 <<< > fio_randwrite_2m_32: 811 <<< > fio_write_2m_128: 833 <<< > fio_write_8m_64: 901 <<< > > Best Regards, > -Dieter > > > On Wed, Aug 29, 2012 at 10:50:12AM +0200, Alexandre DERUMIER wrote: > > Nice results ! > > (can you make same benchmark from a qemu-kvm guest with virtio-driver ? > > I have made some bench some month ago with stephan priebe, and we never be able to have more than 20000iops, with a full ssd 3nodes cluster) > > > > >>How can I set the variables when the Journal data have go to the OSD ? (after X seconds and/or when Y %-full) > > I think you can try to tune these values > > > > filestore max sync interval = 30 > > filestore min sync interval = 29 > > filestore flusher = false > > filestore queue max ops = 10000 > > > > > > > > ----- Mail original ----- > > > > De: "Dieter Kasper" <d.kasper@xxxxxxxxxxxx> > > À: ceph-devel@xxxxxxxxxxxxxxx > > Cc: "Dieter Kasper (KD)" <d.kasper@xxxxxxxxxxxx> > > Envoyé: Mardi 28 Août 2012 19:48:42 > > Objet: RBD performance - tuning hints > > > > Hi, > > > > on my 4-node system (SSD + 10GbE, see bench-config.txt for details) > > I can observe a pretty nice rados bench performance > > (see bench-rados.txt for details): > > > > Bandwidth (MB/sec): 961.710 > > Max bandwidth (MB/sec): 1040 > > Min bandwidth (MB/sec): 772 > > > > > > Also the bandwidth performance generated with > > fio --filename=/dev/rbd1 --direct=1 --rw=$io --bs=$bs --size=2G --iodepth=$threads --ioengine=libaio --runtime=60 --group_reporting --name=file1 --output=fio_${io}_${bs}_${threads} > > > > .... is acceptable, e.g. > > fio_write_4m_16 795 MB/s > > fio_randwrite_8m_128 717 MB/s > > fio_randwrite_8m_16 714 MB/s > > fio_randwrite_2m_32 692 MB/s > > > > > > But, the write IOPS seems to be limited around 19k ... > > RBD 4M 64k (= optimal_io_size) > > fio_randread_512_128 53286 55925 > > fio_randread_4k_128 51110 44382 > > fio_randread_8k_128 30854 29938 > > fio_randwrite_512_128 18888 2386 > > fio_randwrite_512_64 18844 2582 > > fio_randwrite_8k_64 17350 2445 > > (...) > > fio_read_4k_128 10073 53151 > > fio_read_4k_64 9500 39757 > > fio_read_4k_32 9220 23650 > > (...) > > fio_read_4k_16 9122 14322 > > fio_write_4k_128 2190 14306 > > fio_read_8k_32 706 13894 > > fio_write_4k_64 2197 12297 > > fio_write_8k_64 3563 11705 > > fio_write_8k_128 3444 11219 > > > > > > Any hints for tuning the IOPS (read and/or write) would be appreciated. > > > > How can I set the variables when the Journal data have go to the OSD ? (after X seconds and/or when Y %-full) > > > > > > Kind Regards, > > -Dieter > > > > > > > > -- > > > > -- > > > > > > > > > > > > Alexandre D e rumier > > > > Ingénieur Systèmes et Réseaux > > > > > > Fixe : 03 20 68 88 85 > > > > Fax : 03 20 68 90 88 > > > > > > 45 Bvd du Général Leclerc 59100 Roubaix > > 12 rue Marivaux 75002 Paris > > -- > > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > > the body of a message to majordomo@xxxxxxxxxxxxxxx > > More majordomo info at http://vger.kernel.org/majordomo-info.html > > > > > -- > > -- > > > > > > Alexandre D e rumier > > Ingénieur Systèmes et Réseaux > > > Fixe : 03 20 68 88 85 > > Fax : 03 20 68 90 88 > > > 45 Bvd du Général Leclerc 59100 Roubaix > 12 rue Marivaux 75002 Paris > -- -- Alexandre D e rumier Ingénieur Systèmes et Réseaux Fixe : 03 20 68 88 85 Fax : 03 20 68 90 88 45 Bvd du Général Leclerc 59100 Roubaix 12 rue Marivaux 75002 Paris -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html