Re: poor OSD performance using kernel 3.4

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



It would be really nice if somebody from inktank can comment this whole
sitation.

Thanks!

Stefan

Am 29.05.2012 05:54, schrieb Alexandre DERUMIER:
>>> This happens with ext4 or btrfs too. 
> 
> maybe this is related to io scheduler ?
> 
> did you have compared cfq,deadline,noop scheduler ?
> 
> noop should be fast with ssd.
> 
> 
> also what's is your sas/sata controller  ?
> 
> ----- Mail original ----- 
> 
> De: "Stefan Priebe" <s.priebe@xxxxxxxxxxxx> 
> À: "Alexandre DERUMIER" <aderumier@xxxxxxxxx> 
> Cc: ceph-devel@xxxxxxxxxxxxxxx, "Mark Nelson" <mark.nelson@xxxxxxxxxxx> 
> Envoyé: Lundi 28 Mai 2012 21:48:34 
> Objet: Re: poor OSD performance using kernel 3.4 
> 
> Am 28.05.2012 08:52, schrieb Alexandre DERUMIER: 
>>> I think filestore journal parallel works only with btrfs. 
>>> Other filesystem are writeahead. 
>>>> ... you might be right but i can't change ceph's implementation. 
>>
>> See my schema, 
>> I think you see parallel writes, because you see flush write of first wave to disk, in the same time 
>> of second wave write to journal. 
> Yes i fulllý understand and agree - but still this should at least 
> result in a constant bandwidth near max of underlying disk. 
> 
>>>> I totally aggree with you but this is just a test setup AND if you have 
>>>> a big log file to copy let's say 100GB your journal will never be big 
>>>> enough and the speed should never drop to 0MB/s. Also i see the correct 
>>>> behaviour with 3.0.X where the speed is maxed to the underlying device. 
>>>> So i still see no reason that with 3.4 the speed drops to 0MB/s and is 
>>>> mostly 10-20MB/s instead of 130MB/s. 
>>
>> Maybe something is wrong with 3.4, then your disk write more slowly. (xfs bug, sata driver controller bug, ...) 
> 
> This happens with ext4 or btrfs too. 
> 
> Squential write speed to FS is exactly the same under 3.0 and 3.4 using 
> oflag=direct. 
> 
> 3.4: 
> 10000+0 records in 
> 10000+0 records out 
> 10485760000 bytes (10 GB) copied, 41,4899 s, 253 MB/s 
> 
> 3.0: 
> 10000+0 records in 
> 10000+0 records out 
> 10485760000 bytes (10 GB) copied, 40,861 s, 257 MB/s 
> 
>> maybe some local benchmark of your ssd with 3.4 can give some tips ? 
> 
>>>> How many disks (7,2K) do you have by osd ? 
>>>>> One intel 520 SSD per OSD. 
>>
>> I see some benchmark on internet about 150-300MB/s (depend of the blocksize). 
> bench OSD shows around 260MB/s 
> 
> ceph osd tell X bench shows me a speed of 260MB/s under both kernels 
> which corresponds to the dd from above. 
> 
>> Something must be wrong, Doing local benchmark can really help I think. 
>> You can use sysbench-tools 
>> https://github.com/tsuna/sysbench-tools 
>> It make bench compare with nice graphs. 
> Thx hopefully i'll find something. 
> 
> Stefan 
> 
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux