On 8/18/21 9:22 PM, 신희원 / 학생 / 컴퓨터공학부 wrote:
Hi,
I measured the performance of ceph-osd and crimson-osd with same single
core affinity.
I checked IOPS, Latency by rados-bench write , and crimson-osd has lower
performance than ceph-osd about 3 times. (ceph-osd(BlueStore): 228 IOPS,
crimson-osd(AlienStore): 73 IOPS)
-> " $ rados bench -p rbd 10 write --no-cleanup "
Then, crimson-osd's CPU utilization is almost 100%.
I think this is the reason of performance degradation..
Crimson-osd is known for lower CPU consumption than ceph-osd.
I wonder why crimson-osd has more CPU usage in this experiment.
Please let me know how to fix it,
Shin
Those are pretty low numbers in general. Is that on HDD? You are
correct that crimson right now is only using a single reactor core.
That means that tests done with cyanstore will (almost!) always be
limited to ~100% CPU usage. With alienstore you'll still have bluestore
worker threads, but we are often still limited by work being done in the
reactor thread. Here's the most recent performance data we've got
internally (from a ~july build of crimson-osd vs ceph-osd) on NVMe
drives using fio:
https://docs.google.com/spreadsheets/d/1AXj9h0yDc2ztFWuptqcTrNU2Ui3wMyAn6QUft3CPdcc/edit?usp=sharing
The gist of it is that on the read path, crimson+cyanstore is
significantly more efficient than crimson+alienstore and any classic
setup. We are slower in terms of absolute performance, but that's
expected to be the case until the multi-reactor work is done. The
thinking right now is that we probably have some optimization we can do
alienstore and of course bluestore as well (We have ongoing work
there). On the write path things are a little murkier. Cyanstore for
some reason is more efficient with very small and very large datasets
but not the middle size case. alienstore/bluestore and classic memstore
efficiency seems to drop overall as the dataset size grows. In fact
classic memstore is significantly less efficient on the write path than
bluestore is and this isn't the first dataset to show this.
I guess all of this is a round about way of saying that you are testing
very "wild west" code right now. 73 iops is pretty abysmal, but if this
is on HDD you might be the only person that's ever tried
crimson+alienstore+hdd so far. We might be more sensitive to backend
latency on HDD with crimson+alienstore, but that's just a guess.
Mark
ᐧ
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx