Hi there,
just wanted to share some benchmark experience with RBD caching, that I have just (partially) implemented. This is not nicely formated results, just raw numbers to understadn the difference
*** INFRASTRUCTURE:
- 3 hosts with: 12 x 4TB drives, 6 Journals on 1 SSD, 6 journals on second SSD
- 10GB NICs on both Compute and Storage nodes
- 10GB dedicated replication/private CEPH network
- Libvirt 1.2.3
- Qemu 0.12.1.2
- qemu drive-cache=none (set by CloudStack)
*** CEPH SETTINGS (ceph.conf on KVM hosts):
[client]
rbd cache = true
rbd cache size = 67108864 # (64MB)
rbd cache max dirty = 50331648 # (48MB)
rbd cache target dirty = 33554432 # (32MB)
rbd cache max dirty age = 2
rbd cache writethrough until flush = true # For safety reasons
*** NUMBERS (CentOS 6.6 VM - FIO/sysbench tools):
Random write, NO RBD cache: 170 IOPS !!!!
Random write, RBD cache 64MB: 6500 IOPS.
Sequential writes improved from ~ 40 MB/s to 800 MB/s
Will check latency also...and let you know
*** IMPORTANT:
Make sure to have latest VirtIO drivers, because:
- CentOS 6.6, Kernel 2.6.32.x - *RBD caching does not work* (2.6.32 VirtiIO driver does not send flushes properly)
- CentOS 6.6 Kernel 3.10 Elrepo *RBD caching works fine* (new VirtIO drivers sending flushes fine)
I dont know for Windows, but will give you "before" and "after" numbers very soon.
Best,
-- Andrija Panić
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com