Re: Recent ceph.io Performance Blog Posts

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 11/8/22 21:20, Mark Nelson wrote:

2.
    https://ceph.io/en/news/blog/2022/qemu-kvm-tuning/
    <https://ceph.io/en/news/blog/2022/qemu-kvm-tuning/>

You tested network encryption impact on performance. It would be nice to see how OSD encryption (encryption at rest) impacts performance. As far as I can see there is not much public information available on this. However there is one thread with this exact question asked [1]. And it contains an interesting blog post from Cloudlare [2]. I repeated the tests from [2] and could draw the same conclusions: TL;DR: performance is increased a lot and less CPU is used. Some fio 4k write, iodepth=1, performance numbers on a Samsung PM983 3.84 TB drive )Ubuntu 22.04 with HWE kernel, 5.15.0-52-generic, AMD EPYC 7302P 16-Core Processor, C-state pinning, CPU performance mode on, Samsung PM 983 firmware: EDA5702Q):

Unencrypted NVMe:

write: IOPS=63.3k, BW=247MiB/s (259MB/s)(62.6GiB/259207msec); 0 zone resets
    clat (nsec): min=13190, max=56400, avg=15397.89, stdev=1506.45
     lat (nsec): min=13250, max=56940, avg=15462.03, stdev=1507.88


Encrypted (without no_write_workqueue / no_read_workqueue):

write: IOPS=34.8k, BW=136MiB/s (143MB/s)(47.4GiB/357175msec); 0 zone resets
    clat (usec): min=24, max=1221, avg=28.12, stdev= 2.98
     lat (usec): min=24, max=1221, avg=28.37, stdev= 2.99


Encrypted (with no_write_workqueue / no_read_workqueue enabled):

write: IOPS=55.7k, BW=218MiB/s (228MB/s)(57.3GiB/269574msec); 0 zone resets
    clat (nsec): min=15710, max=87090, avg=17550.99, stdev=875.72
     lat (nsec): min=15770, max=87150, avg=17614.82, stdev=876.85

So encryption does have a performance impact, but the added latency compared to the latency Ceph itself adds to (client) IO seems negligible. At least, when the work queues are bypassed, otherwise a lot of CPU seems to be involved (loads of kcryptd threads). And that might hurt max performance on a system that is CPU bound.

From the talk "DATA SECURITY AND STORAGE HARDENING IN ROOK AND CEPH" [3] I understood that > 95% of all Red Hat's customers use OSD encryption. So this might potentially be a big performance gain for all (RH) users of Ceph that use encryption (and have fast SSD / NVMe drives).

I'm currently busy on enabling OSD encryption on a test cluster. As soon as that is finished (will probably take a couple of days) I want test real life impact of the bypass read / write work queue on a Ceph cluster. Adding support for this in Ceph seems pretty trivial, as all that would be required is adding the options to the command that opens the crypt device in ceph-volume [4].


Gr. Stefan

[1]: https://lists.ceph.io/hyperkitty/list/ceph-users@xxxxxxx/thread/44GMO5UGOXDZKFSOQMCPPHYTREUEA3ZI/

[2]: https://blog.cloudflare.com/speeding-up-linux-disk-encryption/

[3]: https://virtual-event-2022.ceph.io/en/community/events/2022/ceph-virtual/

[4]: https://github.com/ceph/ceph/blob/main/src/ceph-volume/ceph_volume/util/encryption.py#L70
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux