Why is librbd1 / librados2 from Firefly 20% slower than the one from dumpling?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Am 02.07.2014 00:51, schrieb Gregory Farnum:
> On Thu, Jun 26, 2014 at 11:49 PM, Stefan Priebe - Profihost AG
> <s.priebe at profihost.ag> wrote:
>> Hi Greg,
>>
>> Am 26.06.2014 02:17, schrieb Gregory Farnum:
>>> Sorry we let this drop; we've all been busy traveling and things.
>>>
>>> There have been a lot of changes to librados between Dumpling and
>>> Firefly, but we have no idea what would have made it slower. Can you
>>> provide more details about how you were running these tests?
>>
>> it's just a normal fio run:
>> fio --ioengine=rbd --bs=4k --name=foo --invalidate=0
>> --readwrite=randwrite --iodepth=32 --rbdname=fio_test2 --pool=teststor
>> --runtime=90 --numjobs=32 --direct=1 --group
>>
>> Running one time with firefly libs and one time with dumpling libs.
>> Traget is always the same pool on a firefly ceph storage.
> 
> What's the backing cluster you're running against? What kind of CPU
> usage do you see with both? 25k IOPS is definitely getting up there,
> but I'd like some guidance about whether we're looking for a reduction
> in parallelism, or an increase in per-op costs, or something else.

Hi Greg,

i don't have that test cluster anymore. It had to go into production
with dumpling.

So i can't tell you.

Sorry.

Stefan

> -Greg
> Software Engineer #42 @ http://inktank.com | http://ceph.com
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo at vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux