SSD Caching Mode Question

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



-----BEGIN PGP SIGNED MESSAGE-----
Hash: SHA256

We are inserting an SSD tier into our very busy cluster and I have a
question regrading writeback and forward modes.

Write back is the "normal" mode for RBD with VMs. When we put the tier
in writeback mode we see objects are being promoted and once the ratio
is reached objects are evicted, this works as expected. When we place
the tier into forward mode, we don't see any objects being evicted to
the base tier when they are written to as described in the manual [1].
Is this a bug? We are running 0.94.5.

Now, I usually like things to work they way they are described in the
manual, however this "bug" is a bit advantageous for us. It appears
that we don't have enough IOPs in the SSD tier to handle the steady
state (we still  have some more SSDs to add in, but it requires
shuffling hardware around). However, when we put the tier into forward
mode, the latency drops and we get much more performance from the Ceph
cluster. In write back we seem to be capped at about 9K IOPs accroding
to ceph -w with spikes up to about 15K. However in forward mode we can
hit 65K IOPs and have a stead state near 30K IOPs. I'm linking two
graphs to show what I'm describing (for some reason the graphs seem to
be half of what is reported by ceph -w). [2][3]

Does the promote/evict logic really add that much latency? It seems
that overall the tier performance can be very good. We are using three
hit sets with 10 minutes per set and all three sets have to have a
read to promote it (we don't want to promote isolated reads). Does
someone have some suggestions from getting the forward like
performance in writeback?

We have 35 1 TB Micron M600 drives ( 26K single thread direct sync 4K
random writes, 43K two thread test, we are already aware of the
potential power loss issue so you don't need to bring that up) in 3x
replication. Our current hot set is about 4.5TB and only shifts by
about 30% over a week's time. We have cache_target_full_ratio set to
0.55 so that we leave a good part of the drive empty for performance.
Also about 90% of our reads are in 10% of the working set and 80% of
our writes are in about 20% of the working set.

[1] http://docs.ceph.com/docs/master/rados/operations/cache-tiering/#removing-a-writeback-cache
[2] http://robert.leblancnet.us/files/performance.png
[3] http://robert.leblancnet.us/files/promote_evict.png


Thanks,
- ----------------
Robert LeBlanc
PGP Fingerprint 79A2 9CA4 6CC4 45DD A904  C70E E654 3BB2 FA62 B9F1
-----BEGIN PGP SIGNATURE-----
Version: Mailvelope v1.2.3
Comment: https://www.mailvelope.com

wsFcBAEBCAAQBQJWS8qbCRDmVDuy+mK58QAADboQAL0tl1ZArL1zPFBf5lYh
xuYQyaWsoaOgdPvlsFhciSrh3VmdTkT9R3O6MZ61VEauKUHmoipE39KejPj3
dQMKKHYc+6VF1MoNoQbeml63jC3DJGBDhPOd+bQ7RE8GBaKM71JaWvvG5bgW
xLAZ7F+37jpHkp/9syrnb0wMxOtZ0xq/iW8Kt3lvSz5Qx6XNx5r78+H9Zr28
OO4xFK8JNfa3JK7RbYU3VeUZCRhhIk/Enb8NdpA0a2cT1meTKfHMDKlOWmT4
qrWIfptWdtADveq6xY2Kj92dFXVnwNfFjoIl4PXTwZtZM1RyAc1gy3qMBADI
BOvn5jdw1PmVYHpY9NH57vpKhn+5o6+FvW95baE5OFJ52NthkVp87LuutnKV
RNyy/cWEe2/Dc9QZdj3eXKjEcL5MYgM+P21THO2e7QQwD6GXnJWnsSTwsQOm
Qs6RqyE9RgdpabdThRzxWIuT8TJmBrDOovEulzFpBN3ZG8bsOrS/5pTmgamI
c8FyddhFgYsPwjMKEDvEbHTIPHx1tZ9hL5fjAwZQeMMCV3LWojAK33a0a602
JfSBj1dhICaULUFQT9f9yhd8/maYNpWogHb/zb3wolegcVP0UckcVxNEUxIf
hxpTvFV93BzUupaprn03Oje2qbSdY++9lZbBfkVChodEprM5oejiT158WBYr
Z3Af
=FP/u
-----END PGP SIGNATURE-----
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux