Re: Performance degredation with upgrade from Octopus to Pacific

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Then highly likely you're bitten by https://tracker.ceph.com/issues/52089

This has been fixed starting 16.2.6. So please update or wait for a bit till 16.2.7 is release which is gonna to happend shortly.


Thanks,

Igor


On 11/1/2021 7:25 PM, Dustin Lagoy wrote:
I am running a cephadm base cluster with all images on 16.2.5.

Thanks for the quick response!
Dustin


On Mon, Nov 01, 2021 at 07:18:38PM +0300, Igor Fedotov wrote:
Hey Dustin,

what Pacific version have you got?


Thanks,

Igor

On 11/1/2021 7:08 PM, Dustin Lagoy wrote:
Hi everyone,

This is my first time posting here, so it's nice to meet you all!

I have a Ceph cluster that was recently upgraded from Octopus to Pacific and now the write performance is noticeably worse. We have an application which continually writes 270MB/s through the RGW which was working fine before the upgrade. Now it struggles to write 170MB/s continuously.

I don't have detailed benchmarks from before the upgrade other that what I just mentioned but I have investigated some since. For background the cluster is two hosts with 10 HDD OSDs each which is under 5% usage overall. The RGW pool is set up with a k=4, m=2 erasure code (across OSDs).

Profiling with `rados bench` to a pool with the same erasure code setup gives similar 170MB/s performance (so about 42MB/s per disk). Profiling a single OSD using `ceph tell osd.0 bench` yields an average of 40MB/s across all disks (which should yield close to 160MB/s at the pool level). Given this it seems to me to be an issue at the OSD level.

I tried setting `bluefs_buffered_io` to false as mentioned elsewhere. This reduced the `%wrqm` reported by iostat (which was previously close to 100%) and gave a slight performance gain (around 50MB/s per OSD), but nothing close to what was seen previously.

Before and after the above change both iostat and iotop report over 100MB/s written to disk while a single `ceph tell osd bench` command is running and iotop shows a near threefold write amplification. With the benchmark writing 1GB to disk iotop shows both the `bstore_kv_sync` and `bstore_kv_final` threads writing about 1GB each and the threads `rocksdb:high0` and `rocksdb:low0` writing a total of 1GB. So 3GB total for the 1GB benchmark.

Looking at the OSD logs during the benchmark it seems rocksdb is compacting several times. I tried adding sharding to one of the OSDs as mentioned in the documentation (with `ceph-bluestore-tool`) but it didn't seem to make a difference.

Does anyone have any idea what may have caused this performance loss? I am happy to post any more logs/detail if it would help.

Thanks!
Dustin

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux