Re: Performance degredation with upgrade from Octopus to Pacific

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks Igor! I'll try to upgrade an report back.

Dustin

On Mon, Nov 01, 2021 at 07:27:11PM +0300, Igor Fedotov wrote:
Then highly likely you're bitten by https://tracker.ceph.com/issues/52089

This has been fixed starting 16.2.6. So please update or wait for a bit
till 16.2.7 is release which is gonna to happend shortly.


Thanks,

Igor


On 11/1/2021 7:25 PM, Dustin Lagoy wrote:
> I am running a cephadm base cluster with all images on 16.2.5.
>
> Thanks for the quick response!
> Dustin
>
>
> On Mon, Nov 01, 2021 at 07:18:38PM +0300, Igor Fedotov wrote:
> Hey Dustin,
>
> what Pacific version have you got?
>
>
> Thanks,
>
> Igor
>
> On 11/1/2021 7:08 PM, Dustin Lagoy wrote:
>> Hi everyone,
>>
>> This is my first time posting here, so it's nice to meet you all!
>>
>> I have a Ceph cluster that was recently upgraded from Octopus to Pacific and now the write performance is noticeably worse. We have an application which continually writes 270MB/s through the RGW which was working fine before the upgrade. Now it struggles to write 170MB/s continuously.
>>
>> I don't have detailed benchmarks from before the upgrade other that what I just mentioned but I have investigated some since. For background the cluster is two hosts with 10 HDD OSDs each which is under 5% usage overall. The RGW pool is set up with a k=4, m=2 erasure code (across OSDs).
>>
>> Profiling with `rados bench` to a pool with the same erasure code setup gives similar 170MB/s performance (so about 42MB/s per disk). Profiling a single OSD using `ceph tell osd.0 bench` yields an average of 40MB/s across all disks (which should yield close to 160MB/s at the pool level). Given this it seems to me to be an issue at the OSD level.
>>
>> I tried setting `bluefs_buffered_io` to false as mentioned elsewhere. This reduced the `%wrqm` reported by iostat (which was previously close to 100%) and gave a slight performance gain (around 50MB/s per OSD), but nothing close to what was seen previously.
>>
>> Before and after the above change both iostat and iotop report over 100MB/s written to disk while a single `ceph tell osd bench` command is running and iotop shows a near threefold write amplification. With the benchmark writing 1GB to disk iotop shows both the `bstore_kv_sync` and `bstore_kv_final` threads writing about 1GB each and the threads `rocksdb:high0` and `rocksdb:low0` writing a total of 1GB. So 3GB total for the 1GB benchmark.
>>
>> Looking at the OSD logs during the benchmark it seems rocksdb is compacting several times. I tried adding sharding to one of the OSDs as mentioned in the documentation (with `ceph-bluestore-tool`) but it didn't seem to make a difference.
>>
>> Does anyone have any idea what may have caused this performance loss? I am happy to post any more logs/detail if it would help.
>>
>> Thanks!
>> Dustin
>>
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux